Artiﬁcial Intelligence and Music: Open Questions of Copyright Law and Engineering Praxis

: The application of artiﬁcial intelligence (AI) to music stretches back many decades, and presents numerous unique opportunities for a variety of uses, such as the recommendation of recorded music from massive commercial archives, or the (semi-)automated creation of music. Due to unparalleled access to music data and effective learning algorithms running on high-powered computational hardware, AI is now producing surprising outcomes in a domain fully entrenched in human creativity—not to mention a revenue source around the globe. These developments call for a close inspection of what is occurring, and consideration of how it is changing and can change our relationship with music for better and for worse. This article looks at AI applied to music from two perspectives: copyright law and engineering praxis. It grounds its discussion in the development and use of a speciﬁc application of AI in music creation, which raises further and unanticipated questions. Most of the questions collected in this article are open as their answers are not yet clear at this time, but they are nonetheless important to consider as AI technologies develop and are applied more widely to music, not to mention other domains centred on human creativity.


Introduction
Artificial intelligence (AI) is a well-established discipline of computer science focused on making computers perform tasks that would normally require human intelligence (Russell and Norvig 1995). Due to the convergence of massive data availability, computational resources and novel deep-learning-based architectures, the machine learning (ML) sub-field of AI has experienced major breakthroughs over the past decade (Goodfellow et al. 2018). Such great progress has been made that AI technology is now a major driver of global business and investment. 1 AI is being deployed in varied practical scenarios, from machine translation to medical diagnosis, detecting fraudulent credit card use, transcribing speech, summarizing sporting events and financial reports, targeted advertising on social media, and autonomous vehicles. The combination of AI and robotics is also widely transforming manufacturing (Economics 2019). Real-world applications of AI have generated controversy as well, such as tracking people using biometric indicators (e.g., face, speech, gait), prediction of criminal 1 For instance, see the number of articles appearing like A. Moore's "When AI Becomes an Everyday Technology", Harvard Business Review, 7 June 2019, https://hbr.org/2019/06/when-ai-becomes-an-everyday-technology.
data. An online implementation 18 offers users the functionality of generating transcriptions with the click of a button. Figure 1 shows the parameters available to the user, and a notated transcription generated by the AI. Users can submit the transcription to a growing collection of such tunes at another website. 19 Material generated by folkrnn models has been used in variety of ways, 20 including an album situated within the style of Irish traditional music . 21 4 4 Figure 1. The menu of http://folkrnn.org (left) gives the user parameters with which to direct the music generation process of a selected AI model. Here, the user has selected the model "thesession.org (w/ :| |:)", a random meter and mode, a temperature of 1, and a random number seed of 793470. Clicking "Compose" begins the generation process, which creates the transcription shown notated at right. Figure created by authors.
The experience of developing and using folkrnn models brings focus to the numerous questions identified above, and also motivates other questions. Who owns the rights to the material generated by folkrnn models, and to any music derived from it? What implications are there that folkrnn models are trained on a collective source of music transcriptions, some of which may be derived from copyright-protected work? How might folkrnn-and tools like it-impact music creators and music communities, traditional or otherwise? What kinds of misuses of folkrnn are possible, and how can its developers guard against these? How is the value of music affected by AI system that can generate and synthesize millions of "folk music" recordings without the direct input of human musicians? Should a person who uses material generated by folkrnn models in the composition process specify where it came from, i.e., that it was not entirely their own creation? Should music created in such ways, with the "input of AI", be considered inferior to music created without?
These questions motivate taking a closer look at what is actually occurring in the application of AI to music from two different and complementary perspectives: copyright law and engineering praxis. In Section 2, we look specifically at copyright law, and how it might address some of the questions surrounding material generated by music AI. The legal perspective is important because the use of AI in music creation is posing novel challenges to intellectual property law, moral rights, and the protection of creative human production. In Section 3, we look at engineering praxis in AI and ML, and some of the questions arising from their research and development methodologies. This perspective shifts the focus in engineering from developing music AI, and the incentives of doing so, to surveying the impact of such technology and its development, intended or not, and the implicit assumptions made in performing such work. Section 4 returns to folkrnn to discuss aspects of these two perspectives in a more grounded way. The conclusion looks to the future of AI in music, and speculates what it may mean for copyright law and engineering praxis. Overall, this article contributes new perspectives to the interdisciplinary domain of music AI development, which is currently heavily focused on how to create such technology. There are no easy answers to many of the questions raised in this article, but they must continue to be identified and formalized to assist their future discussion.

Copyright Law Perspective
As an artistic work, music may be protected by copyright and related or neighbouring rights. While copyright protects original works, i.e., the musical composition or the lyrics, in the case of music, neighbouring rights relate to the performance or interpretation, or to the fixation made for the sound recording. Thus in addition to the composer, the writer, musicians, singers and phonogram 22 producers also benefit from certain protection.
Focusing on copyright, some countries such as the UK, South Africa, Hong Kong, India, Ireland, and New Zealand have envisaged protection for computed-generated works granted to the person by whom the arrangements necessary for the creation of the work have been undertaken. Thus, in the UK, computer-generated works are defined as works "generated by computer in circumstances such that there is no human author of the work". 23 Note these provisions leave room for ownership being allocated either to the programmer or to the user of the computer program.
In relation to other countries where no specific regime exists (the majority in Europe), it has been questioned whether AI-generated works attract copyright protection. Continental copyright legislation is very much dependent on human-centred concepts, both with regards to the beneficiary of protection (i.e., the author), the conditions for protection (e.g., originality), and the rights granted (economic, but also moral rights). According to the Court of Justice of the European Union (CJEU), a work is considered original when it is the expression of the author's own intellectual creation and his/her free creative choices, the author's personality, or the author's personal touch. 24 In light of this, a number of scholars conclude that under present law, autonomously AI-generated works might not be eligible for copyright protection, e.g., Buning (2018); Deltorn and Macrez (2019); Guadamuz (2017); Lauber-Rönsberg and Hetmank (2019); Michaux (2018); Ramalho (2017). 25 Regardless, humans can still have an important involvement in creating music, even if assisted by an AI system. Originality, as a precondition for copyright-ability, may be concluded whenever there is a significant creative human contribution to the resulting output. However, it cannot be disregarded that in the future, AI might be able to autonomously compose songs with minimal or insubstantial human intervention. 26 Authorship recognition may require an analysis of the operation of the systems and the role of the different actors involved in the process (e.g., the developer, the trainer or the user). 27 These cases are the ones that challenge EU copyright law. 22 Art. 2 of the WIPO Performances and Phonograms Treaty (WPPT) defines phonogram as "the fixation of the sounds of a performance or of other sounds, or of a representation of sounds, other than in the form of a fixation incorporated in a cinematographic or other audiovisual work". 23 2019, Vol. 14, No. 8. 26 In any case, humans would be needed, of course, for creating the technology. This does not mean that the generated music necessarily reflects the personality of those developers. 27 For a detailed discussion of possible authors in relation to music generated by different AI systems see Deltorn and Macrez (2019).
Is the lack of copyright protection of AI-generated results adequate from a policy point of view? The response to this question would require further legal and socio-economic analysis. 28 In case a certain level of protection is considered as necessary, adjustments may be needed to the existing framework to either amend the existing copyright laws or to pass new sui generis rights targeting AI-generated products. One of the obstacles to put in place a different system of protection for AI-generated products would be to elucidate when a work is or is not AI-generated and to what extent.
Another aspect of relevance is the use of copyright works in datasets used for training AI systems. Training state-of-the-art ML models requires large amounts of data, i.e., pre-existing music, such as scores, lyrics and/or audio recordings. When the music used to train a system is protected by copyright, 29 permission from the rightholders is required unless an exception applies. In the EU a limited number of exceptions exist. Copyright exceptions traditionally concern a qualified purpose, e.g., quotation, parody, teaching, research, and news reporting. The new European directive on Copyright in the Digital Single Market 30 has introduced two exceptions on text and data mining 31 that could greatly benefit AI developers. Subject to certain conditions, the first exception 32 would allow text and data mining for purposes of scientific research provided the researcher has lawful access to the work. This exception is mandatory and cannot be derogated by contract. The second exception 33 would authorize text and data mining of lawfully accessible works and other subject matter beyond scientific purposes, except if rightholders have expressly restricted the use. 34 The new provisions could then allow certain reproductions required for automated computational analysis of music hosted on digital repositories if all the required conditions are met. All in all, it must be retained that copyright is about use and not about access; the exceptions can only be invoked in case users have legal access to the work, or in the above mentioned example to the digital repository. Furthermore, the European member states still need to implement the directive into national law.
Other questions arise from behaviors of music AI systems. The accidental reproduction of copyright-protected work in music generated by an AI system may also raise liability questions. In case a piece of music reproduces a pre-existing work or a phonogram, or part of those, the authorization of the relevant rightholders is normally required. In a recent case dealing with the sampling of 2 seconds in a music composition, the CJEU concluded that taking a sound sample of a phonogram for the purposes of including that sample in another phonogram, even if very short, requires the rightholders' authorization, unless that sample is included in a modified form unrecognizable to the ear, or the use is made under a copyright exception. 35 Last but not least, another question concerns the allocation of responsibility for such infringement, which could relate to both the engineer and the user. How should a music AI engineer safeguard against such behaviors?

Engineering Praxis Perspective
Since engineers bring new technologies into the world, it can be argued that they share the responsibility for the resulting outcomes, positive and negative, intended and unintended. 28 The European Commission has recently published a call for a study that will assess whether the current IPR framework is fit-for-purpose for AI-generated works/inventions: https://ec.europa.eu/digital-single-market/en/news/trends-anddevelopments-artificial-intelligence-challenges-intellectual-property-rights. 29 In Europe copyright lasts for 70 years from the death of the author. The duration of neighbouring rights is shorter (50 years) except for published phonograms. The use of works in the public domain is free. Data mining is defined in Article 2 of the directive as "any automated analytical technique aimed at analysing text and data in digital form in order to generate information which includes but is not limited to patterns, trends and correlations". 32 ibid., Article 3. 33 ibid., Article 4. 34 A rightholder can restrict use, e.g., explicitly using methods to block data mining is a way to opt out from the exception in case of material made publicly available online. 35 Judgment of 29 July 2019, Pelham, C-476/17, ECLI:EU:C:2019:624.
The integration of AI into societal and political spheres can appear as fostering objectivity, but the data-centric nature of ML can unintentionally perpetuate discrimination by encoding existing social biases learned from real data used by an engineer for training (Barocas and Selbst 2016). Even if an engineer has good intentions, their uncritical treatment of data and their resulting systems might become instrumental in producing great social and economic harm. No matter how complex a system might be, engineers must study, evaluate and document its working principles and limitations and make users and others impacted by such a system aware of them. Technology is not ethically neutral (Dusek 2006), which motivates design procedures that are guided by clear ethical principles (Hand 2018; The IEEE Global Initiative on Ethics of Autonomous and Intelligent Systems 2017), as well as employing continuous evaluation to mitigate harmful impact. Bryson and Winfield (2017) provide a brief description of and motivations for standardizing the ethical design of AI such that its development benefits humanity. It introduces the standardization efforts of The IEEE Global Initiative on Ethics of Autonomous and Intelligent Systems (2017), and discusses in particular standardizing transparency in autonomous systems. The development and implementation of these practices are still ongoing.
When evaluating the impact of AI, engineers typically work from the narrow perspective of AI performance in limited and controlled contexts. A critical but often overlooked question is what defines a particular system as "good" or "bad" with respect to a use or user, e.g., a music listener, performer, composer or business. Defining meaningful evaluation strategies is difficult for many applications in the music domain, due to the subjectivity of music judgment and preference. This makes evaluation one of the main research challenges in the discipline of music information retrieval (Schedl et al. 2014;Sturm 2016). Engineers by and large focus on performance metrics like accuracy, precision, and reliability when developing data-driven AI techniques, and then infer from performance how useful or harmless a system is. Evaluation can instead be framed in the context of specific applications and their users, but this is rarely done (Schedl et al. 2013).
Broader perspectives are being advocated for by the developing research fields "fair, accountable and transparent machine learning" (FAT/ML), 36 and "explainable AI" or "interpretable AI". These fields argue for richer notions of evaluating AI performance and use, ones that actually reflect usefulness and real impact. Definitions of these concepts are numerous, and represents a track of research in itself. In one sense, an AI system is considered to be fair if it does not discriminate against certain individuals based on a set of protected attributes like gender, race, and age. 37 Group fairness has been proposed as a category of algorithmic fairness (Kleinberg et al. 2017), which consists of more than 21 metrics to measure discrimination, some of which are contradictory. Transparency, or a clear explanation of where and how AI has been used to make some decision or product, has been also argued to be a relevant evaluation criteria for AI development and application. A related concept is interpretability (Doshi-Velez and Kim 2017) or explainability (Mittelstadt et al. 2019): the ability to explain, in an understandable way, how an AI behaves a particular way, or has arrived at a given result. Metcalf (2018) underlines the importance of distinguishing between bias and fairness. The former is a technical feature of statistical models, and the latter is related to human values that comprise social biases and prejudices about a group of people. Bias and unfairness may occur within an AI system without necessarily impacting its performance in some aspects, but can impact metrics reflecting the values of other participants.
These concepts not only promote the measurement of AI systems in ways that are more relevant to their application, but also motivate new methodologies for engineering, and for addressing the general lack of auditing tools to evaluate the social impact of AI. 38 For instance, engineers can use interpretability to find and correct causes of discrimination in an AI system, to better explain its 36 See the FAT/ML workshop: https://www.fatml.org. 37 "Intended" discrimination is commonly known in ML as disparate treatment, while "unintended" as disparate impact. 38 See Kate Crawford (2017), "Artificial intelligence with very real biases" https://www.wsj.com/articles/artificialintelligencewith-very-real-biases-1508252717. errors and improve operation, and to perform causal inference (Zeng et al. 2017). One can also see potential uses of such approaches to addressing some questions of copyright law in music AI raised in Section 2. For instance, explainability can help address questions about how specific music material was generated by a system. Explainability can also improve the deployment of AI systems in the real world, where sample bias and concept drift are common. 39 The engineering of AI systems can benefit by working in transparent ways as well, e.g., clarifying the reasons why specific metrics are used and the societal values which underlie them (Glymour and Herington 2019;Kilbertus et al. 2017); describing the processes of data collection and use, which may result in discrimination (Hand 2018); even improving the composition of a research team, where a lack of team diversity can compound the impact of the problems mentioned above and reinforce blind spots (Ruiz et al. 2002). Working in these ways can help address issues related to the socio-economic impact of AI systems, define new evaluation principles centred around these issues, and prevent and mitigate negative effects, e.g., see Benthall and Haynes (2019). Discussions about bias, fairness and transparency in the development and application of AI to music are just beginning, e.g., Holzapfel et al. (2018). 40 Music creation is a part of a landscape in which various AI-driven tools are deployed and used for different purposes, e.g., composing, mixing, and streaming. For music streaming services, AI can make considerable positive and negative impact on music recommendation. AI can be a core technology of music recommendation systems designed to provide users with relevant recommendations, and thus, can influence music consumption and listening behaviour at large scales. However, such approaches can also incentivise unscrupulous behaviours that take advantage of artists and listeners (Eriksson et al. 2018). 41 For instance, using AI music generation can greatly increase the revenue of a music streaming business, whereby in-house synthetic "artists" are promoted on playlists thus reducing the revenue distributed to human artists whose music is also streamed. In such cases, artists and listeners might be informed about the way AI systems work for music recommendation (Aguiar and Waldfogel 2018). It could also be mandated that users be informed about the ways in which AI was used to create the music they are listening to, or even opt-out of hearing AI-generated music altogether. Such transparency can serve to empower artists and listeners to challenge AI systems; however, these require defining or clarifying the degree to which AI is involved in the process. It is tempting to try and enumerate the involvement of AI in various stages of music creation, e.g., composition, performance, production. For each stage, one might indicate the degree of AI involvement with a numerical score. A more fine grained analysis could describe exactly how AI was involved, e.g., the design of a microphone, the creation of an impulse response, the synthesis of a particular voice, or the use of score following during performance. Consumers could then use such information to make more informed choices, in the same way that labeling food ingredients is used to identify nutritional value or potentially deadly reactions due to allergies.
This approach to transparency, however, faces problems immediately. First, decomposing music creation into stages is not so clear in many instances, and also introduces its own biases, since these stages are not always separable and do not even exist in all music. Second, designating which aspects are more and which are less central to the value consumers attach to music can be debatable. Consider, for instance, an AI application that automatically advances a music score for performing players: certainly useful AI technology, but is it material to the musical value of the recording? Do consumers themselves have enough knowledge about music and its production to evaluate the relevance of such information? Third, an accurate tally of the involvement of AI will also require extensive and sometimes intrusive documentation of all the software and hardware used in the process and its relevance to the finished product-which has the potential to become unwieldy and a hindrance to the creative process, and could even harm artists who do not have institutional backing. Requiring a young, aspiring artist to provide full and accurate documentation of the AI they have used as a condition for making their music available to audiences could be an insurmountable barrier. Investigating the provenance of each product (hardware and software) for traces of AI is perhaps not something many artists would be able or willing to do. This is not to say that it is too difficult to inform consumers about the way AI has been involved in the creation of a particular music experience, but certainly, research must be conducted to establish exactly the harms that can result if such information is not provided to consumers and in what contexts. In the case of a person with a life-threatening allergy to a particular food ingredient, one benefit of labeling food with ingredients is clear: the avoidance of death. Such a benefit is not yet clear in the case of identifying the involvement of AI in music creation. Furthermore, inundating consumers with information but expecting them to ascertain its relevance to them may not in the end be conducive to informed choice.

Implications of Copyright Law and Engineering Praxis for folkrnn
We now turn back to the example of folkrnn presented in the Introduction and examine it from the perspectives of copyright law and engineering praxis.

Legal Perspectives of folkrnn
The data used to create the model generating the music transcription in Figure 1 comes from a crowd-sourced online collection comprised of both pieces composed by people unknown (belonging to the Commons), and pieces with known composers, including newly composed pieces. 42 For over 17 years, thousands of users of that website have contributed over 23,000 transcriptions of music, many of which are traditional Irish, English, and Scottish dance tunes. Should any of these pieces be protected by copyright or the database by the database makers' sui generis right, the permission of the corresponding rightholders may be required for some of the copies needed to make use of the repository for training the system. This is where the new exceptions on text and data mining for the purpose of research, mentioned in Section 2, become highly relevant. Some of the transcriptions used for building the models are based on commercial recordings, but the data are hand-written symbolic transcriptions and not audio samples. Owners of those commercial recordings may ask for such transcriptions to be removed from the dataset.
Another question concerns the possibility that the generated material is plagiarising existing tunes. The aim of folkrnn (and similar AI systems) is to create transcriptions that are similar to the training data but novel (a machine composer). So far, folkrnn models have been used to generate over 100,000 new transcriptions, 43 some of which have been rendered with synthetic musicians (MIDI), or by human musicians-most notably in the music album "Let's Have Another Gan Ainm" . 44 Some of the generated transcriptions have also been posted by users of folkrnn to the website http://themachinefolksession.org, which is itself a crowd-sourced collection of "machine folk" transcriptions. A search of the training data in Figure 1 shows that the generated melody is new, although small parts of it match a few existing transcriptions. For instance, the first bar of the second part matches a bar of a transcription from the thesession.org website. Considering that similarity at that level is extremely common in this type of music, arguing for infringement of copyright seems rather remote. In the case of generated material appearing on the album , plagiarism was not found by using pattern matching, though we must stress that "recognisable similarity" to a human listener (the legal and common sense standard) is not the same.
Returning to the generated music in Figure 1, in light of the explanations given in Section 2, it has to be assessed what is the contribution of the user who is selecting the available parameters and clicking the "compose" button. Is it enough to consider the work original in the sense of European copyright law? On the one hand, the answer is simple: folkrnn only produces a symbolic transcription that needs significant human intervention (e.g., editing, performance and production) to become plausible as music. In that view, folkrnn is a piece of music technology, like the violin or a music editing software. A user that pays for those can use them as they wish. In practice, there is nothing to prevent a user who generated the transcription from claiming copyright of that tune. Though if that copyright was challenged, the users would have a hard time arguing that the few clicks that led to the creation of the score is the expression of their own intellectual creation, their free creative choices, personality, or personal touch.
On the other hand, we may argue that on the basis of contracts or licenses, those who implemented the system and are paying for its maintenance online can claim ownership of the generated transcription, all others the system generates, and/or any derivative works. A folkrnn model is not like word processing software, providing an environment in which content can be created, but the system actually generates content. This seems to be the practice by some commercial websites that are licensing the rights of the music generated by the software even when in some cases, users may have actively contributed to the final input. It must be recalled that if a work is not considered as protected by copyright, the enforcement of the relevant clauses cannot be justified on the basis of copyright law but on the basis of contract law.
Applying AI within the domain of folk music presents an interesting case for evaluating the impact of the technology on music creators and communities and the ways in which AI can be useful or detrimental to musicians. The website http://folkrnn.org is not a generator of finished products, but provides "raw material" to create music. By working with practitioners, both within and outside the folk tradition, Ben-Tal (2017, 2018);  demonstrate that musicians can find interesting uses for that tool. For instance, the performers who worked on the "machine folk" album  found value in learning new tunes and adding their own stamp on them. As one of them remarked, "I particularly fell in love with the waltz 45 and play it regularly and it gets a great response. I just want to do the piece justice by playing it as musically as possible". This illustrates that in addition to commercial ownership and monetary value, there is a personal and social dimension. The musicians took a raw output from folkrnn and added their own imprint in editing, arranging, and performing. They feel musical ownership of the result with the responsibility to communicate it effectively to the audience. This provides validation from both audience and fellow performers (a distinction that is often blurred in Irish traditional music sessions).
By looking at folkrnn and its use and development from the perspective of copyright law, we have provided shades of answers to some of the questions but much has yet to be clarified. The folkrnn models are a potential source of an unlimited amount of material that is stylistically similar to existing music, but efforts should be taken to ensure that generated material does not plagiarise existing music. There is no trivial solution to this challenge. If the model is to plagiarise, it is most likely to be something from the training data, but it could also accidentally plagiarise some other music. There is not at the moment a general music plagiarism detection equivalent to text matching. There are also no clear standards or conventions about what constitutes sufficient alterations to be considered novel. It is still not totally clear if and how works generated by AI systems will be protected by copyright. This question is bound up with our notions of creativity, which is notoriously ill defined 45 The first Gan Ainm on track 3, Gan Ainm, Gan Ainm, Gan Ainm. (Deliège and Wiggins 2006). Advances in music AI technology are likely to provide further challenges to existing legal definitions and social conventions. Furthermore, the use of data for training AI systems may require permission of relevant copyright holders, which can involve huge transaction costs. While folkrnn as a product of research may be exempt from such restrictions, companies building and selling similar tools will not be, at least until new exceptions in the recent European Directive on Copyright in the Digital Single Market are implemented at the national level.

Engineering Praxis Perspectives of folkrnn
We now return to the concepts and questions discussed in Section 3 with reference to folkrnn. First, what makes a folkrnn model "good" or "bad"? Essentially, such a model computes probabilities of symbolic sequences, and so is trained to give a high probability to such sequences from its training dataset. While this criteria is well-defined and facilitates computation, its relevance to the use of such models for music is highly limited. Hence, the developers of folkrnn devised additional evaluation methods bringing music practitioners into the loop. This has involved hiring professional musicians to inspect material generated by folkrnn models -musicians who work both inside and outside the traditions encompassed in the training data Ben-Tal (2017, 2018); . Such expert elicitation in fact led to further questions relating to the value systems of traditional music and its participants; the right of taking a dataset collected for one purpose and using it for a completely different and unanticipated purpose; and in general how, as outsiders to the given traditions, to approach the development and presentation of such systems. These new insights altered the course of further research, and changed the way this project is presented both in academic publications and to the general public. Working closely with expert musicians also led to work on the interpretability of the system (Sturm 2018). Understanding how musicians use an AI system, what they find lacking in it, and their views on the creativity or creative potential they see in the system is a valuable resource to advance computational creativity research. As Agres et al. (2016) argue, an important difference between a generative system and a truly creative one is the integration of self-reflection and the ability to change behaviour accordingly. This is a case where advancing such research is aligned with the social demand that AI systems be explicable.
In addition to what makes a folkrnn model "good" or "bad", the potential harm of such systems should be discussed. Expert elicitation provided insights here as well. The concerns that were heard from musicians of the folkrnn album  were not focused directly at AI displacing creative work, but at disrupting historical and personal links. One of the musicians remarked, "There are so many existing traditional tunes that it would be difficult to devalue other tunes. I would say that it does devalue the tradition from the point of view that the tradition is based on human interaction and history". Similar sentiment was voiced by a user on thesession.org worrying that newcomers will learn these tunes thinking they are the real thing. 46 At the same time musicians recognise the potential of AI to enrich the practice. From amateur musicians who see this as a step towards producing their own tunes to the possibility of the system discovering new possibilities, one of the musicians remarked, "what would it be like to hear an incredible tune which sounded like nothing anyone had imagined before? That would add value, and if not accepted by the tradition, would be the start of a new tradition". However, others in the folk music community have made clear that the very idea of involving AI in folk music embodies a clash of values. 47 The development and use of folkrnn has by and large been open and transparent. The code used for creating the models, as well as the processed datasets, have been publicly available since the project was initiated. 48 The album created from material generated by folkrnn is clearly marked on its host webpage 49 as coming from a music AI; and the attendant technical report  provides further details, including identifying the original material generated by the AI and the changes made by a human expert. The 34 volumes of 100,000 transcriptions generated by folkrnn 50 are also identified as being products of music AI.
If the further development of folkrnn results in an AI that generates the most plausible transcriptions not easily differentiated from the training data, what does that mean for Irish traditional music? In which ways does this impact the experience of folk music? One must question how much danger any such computer model really represents for as strong and living of a tradition as Irish traditional music-and especially one that is focused on performative aspects and has clear social functions (Fairbairn 1993). Though folkrnn models have so far been created from data collected from Irish and Swedish traditional music, they could be further developed for other kinds of music, which may lead to harm. What happens in such a case? These questions have yet to be answered, but they must be addressed continually, and with the collaboration of practitioners as such research progresses (Holzapfel et al. 2018).

Conclusions
Technology is a double-edged sword, with benefits and detriments that deserve to be critically analysed as it is developed, applied, improved and retired. General-purpose technologies such as AI that have the potential to affect economies and societies require study from broad and diverse perspectives, with considerations of questions that touch on law, economy, ecology, and philosophy (Bostrom and Yudkowsky 2014;Craglia 2018;Dusek 2006;Holzapfel et al. 2018). Relevant initiatives in ML research are centred on developing AI systems that are not just accurate, but also fair, accountable, transparent and interpretable. Such work moves beyond describing just the engineering of technologies to examining their use and misuse, considering their impact, and how such dimensions should be reflected in engineering praxis. This article aims to contribute to this discussion specifically for AI developed and applied to music.
We have looked at two broad perspectives, both timely and motivated by a particular music AI application. The first is from the perspective of copyright law, which arises from questions about the ownership of material generated by AI systems that may have themselves been trained on copyright protected material. Many questions have been identified, most of which cannot be answered at this time due to a lack of precedent; but what is certainly clear is the disruption AI music generation will cause to legal and societal norms as it improves. The second perspective is engineering praxis, which arises from concerns as to the impact of AI technology on music ecosystems and the ways in which such engineers work. Many questions are raised here as well that cannot be answered, but what is clear is that music AI should be developed and used in consultation with music practitioners (Holzapfel et al. 2018;.
The impact of music AI on career trajectories and opportunities present risks not just to job seekers, but are likely to have adverse effects on innovation in the music-ecosystem. The entry-level jobs in recording studios and small media companies that are likely to be replaced by music AI allow aspiring musicians to acquire skills and networking opportunities for embarking on their career. Adapting education curricula to include more sophisticated skills and ways of working alongside and with the support of music AI tools is only part of the answer, and temporary at that, with further technological development. Developers of the technology and the managers who decide on deploying it should consider the longer term effects of their decisions, and ways of mitigating adverse results. At the same 48 See https://github.com/IraKorshunova/folk-rnn, https://github.com/victorwegeborn/folk-rnn, https://github.com/ tobyspark/folk-rnn-webapp. 49 See footnote 21. 50 Available at https://highnoongmt.wordpress.com/2018/01/05/volumes-1-20-of-folk-rnn-v1-transcriptions. time, the current legal framework may need adjustment to best facilitate creative activity and artistic innovation in such an evolving ecosystem.
It might be illustrative to consider a possible endpoint of music AI development: a "complete" music AI, a kind of "musical holodeck" able to generate any possible music, archiving all recorded music (scores, audio and video recordings, images, and other data about the music and its performers, etc.). It knows and understands what music is and how it works, and can not only retrieve all music fulfilling the requests of any listener, but also create any possible music in an instant. A subscriber of this system can ask to hear Led Zeppelin performing a particular ABBA song. Composer Ben-Tal can hear his piano music performed by a long-dead pianist or the Berlin Philharmonic. The system provides any subscriber with limitless access to individualised musical experiences. Will there still be a need for recording engineers, producers, and performers? Any subscriber can now be a composer of their own soundtrack. There is no need to distribute music because all music is (re)created locally. Would such a system divorce music from musical "works"? A system that responds dynamically to the wishes of a subscriber will not need to be centred on structured stretches of organised sound. Music could become an amorphous soundtrack to one's life -delivering on Erik Satie's vision of "Furniture Music": music that is heard but not listened to; music that forms part of everyday life like the objects people have in their home (Templier 1969).
Such a future clearly requires a careful assessment of copyright rules. If a subscriber hears something they like, they can ask to hear something "like it" and the AI should be able to produce a new instance that meets their demands without outright copying. What should then be the approach to the ownership of AI-generated music? We would have to consider the purpose of awarding copyright. Generally speaking, copyright (and related rights) aims to provide creative people (and investors) with a reward in the form of an exclusive, quasi monopolistic, right. In droit d'auteur countries, copyright also reveals an intimate link with the personality of the author. Returning to our music AI, its musical potential is immense and it is likely that most subscribers will only ever discover a tiny sliver of that potential. It is also likely that some subscribers will have the inclination and the aptitude to be much more inventive in their interaction with the system and produce truly unique results through it. How can we enable such subscribers to dedicate their time and effort towards this and have means of sharing the fruits of their intellectual labour with others? Is there any right the designer of such an AI should retain? What would be the impact of maintaining the status quo, or of passing new rights in the market of human created and/or AI generated works? As our survey of copyright law indicates, technological developments challenge established norms. The distant future sketched above suggests that there is a need for fundamental re-thinking in this area.
Of course, the inevitability of such a music AI should be taken with a grain of salt. Human creativity can surprise in its ability to incorporate new technologies. For instance, the turntable was designed to facilitate playback at home but has become a performance tool in ways that were not intended by its developers or manufacturer. It is likely that any music AI tools will open opportunities for creative but unintended uses. Funding: This research received no external funding.

Conflicts of Interest:
The authors declare no conflict of interest. The views expressed in the article do not imply a policy position of the European Commission.