Whether or not the edit is marked as minor (no longer in use) (minor_edit ) | false |
Edit count of the user (user_editcount ) | 732 |
Name of the user account (user_name ) | 'Darwin Naz' |
Age of the user account (user_age ) | 9530988 |
Groups (including implicit) the user is in (user_groups ) | [
0 => 'extendedconfirmed',
1 => '*',
2 => 'user',
3 => 'autoconfirmed'
] |
Rights that the user has (user_rights ) | [
0 => 'extendedconfirmed',
1 => 'createaccount',
2 => 'read',
3 => 'edit',
4 => 'createtalk',
5 => 'writeapi',
6 => 'viewmywatchlist',
7 => 'editmywatchlist',
8 => 'viewmyprivateinfo',
9 => 'editmyprivateinfo',
10 => 'editmyoptions',
11 => 'abusefilter-log-detail',
12 => 'centralauth-merge',
13 => 'abusefilter-view',
14 => 'abusefilter-log',
15 => 'vipsscaler-test',
16 => 'collectionsaveasuserpage',
17 => 'reupload-own',
18 => 'move-rootuserpages',
19 => 'move-categorypages',
20 => 'createpage',
21 => 'minoredit',
22 => 'editmyusercss',
23 => 'editmyuserjson',
24 => 'editmyuserjs',
25 => 'purge',
26 => 'sendemail',
27 => 'applychangetags',
28 => 'spamblacklistlog',
29 => 'mwoauthmanagemygrants',
30 => 'reupload',
31 => 'upload',
32 => 'move',
33 => 'collectionsaveascommunitypage',
34 => 'autoconfirmed',
35 => 'editsemiprotected',
36 => 'movestable',
37 => 'autoreview',
38 => 'skipcaptcha',
39 => 'transcode-reset'
] |
Whether the user is editing from mobile app (user_app ) | false |
Whether or not a user is editing through the mobile interface (user_mobile ) | false |
Page ID (page_id ) | 21659435 |
Page namespace (page_namespace ) | 0 |
Page title without namespace (page_title ) | 'Music and artificial intelligence' |
Full page title (page_prefixedtitle ) | 'Music and artificial intelligence' |
Action (action ) | 'edit' |
Edit summary/reason (summary ) | 'intro expansion' |
Old content model (old_content_model ) | 'wikitext' |
New content model (new_content_model ) | 'wikitext' |
Old page wikitext, before the edit (old_wikitext ) | '{{multiple issues|
{{cleanup|date=December 2010}}
{{advert|date=December 2010}}
}}
Research in [[artificial intelligence (AI)]] is known to have impacted [[medical diagnosis]], [[Algorithmic trading|stock trading]], [[robot control]], and several other fields. Perhaps less popular is the contribution of AI in the field of [[music]]. Nevertheless, artificial intelligence and music (AIM) has, for a long time, been a common subject in several conferences and workshops, including the [[International Computer Music Conference]], the Computing Society Conference <ref>[http://www-rcf.usc.edu/~echew/INFORMS/ics2005.html INFORMS Computing Society Conference: Annapolis: Music, Computation and AI] {{webarchive|url=https://archive.is/20120630221600/http://www-rcf.usc.edu/~echew/INFORMS/ics2005.html |date=2012-06-30 }}. Rcf.usc.edu. Retrieved on 2010-12-22.</ref> and the [[International Joint Conference on Artificial Intelligence]]. In fact, the first International Computer Music Conference was the ICMC 1974, Michigan State University, East Lansing, USA <ref>[http://www.computermusic.org/page/23/ International Computer Music Association - ICMC]. Computermusic.org (2010-11-15). Retrieved on 2010-12-22.</ref>
Current research includes the application of AI in music composition, performance, theory and digital sound processing.<ref>[http://www.informaworld.com/smpp/content~db=all~content=a714908257?words=artificial*|intelligence*|music* Research in Music and Artificial Intelligence - Ergonomics Abstracts]. Informaworld.com. Retrieved on 2010-12-22.</ref> Several music software applications have been developed that use AI to produce music<ref>{{cite journal| title = A Functional Taxonomy of Music Generation Systems
| author = D. Herremans, C.H., Chuan, E. Chew
| year = 2017
| journal = ACM Computing Surveys
| volume = 50
| issue = 5
| doi = 10.1109/TAFFC.2017.2737984
| pages = 69:1-30}}</ref>. A few examples are included below. Note that there are many that are still being developed.
==History==
In 1960, Russian researcher R.Kh.Zaripov published worldwide first paper on algorithmic music composing using the "[[Ural (computer)|Ural-1]]" computer.<ref>{{cite journal|last=Zaripov|first=R.Kh.|title=Об алгоритмическом описании процесса сочинения музыки (On algorithmic description of process of music composition)|journal=[[Proceedings of the USSR Academy of Sciences]]|year=1960|volume=132|issue=6}}</ref>
In 1965, inventor Ray Kurzweil premiered a piano piece created by a computer that was capable of pattern recognition in various compositions. The computer was then able to analyze and use these patterns to create novel melodies. The computer was debuted on Steve Allen's I've Got a Secret program, and stumped the hosts until film star Henry Morgan guessed Ray's secret.<ref>http://www.kurzweiltech.com/raybio.html</ref>
==Software applications==
===Interactive scores===
Multimedia Scenarios in interactive scores are represented by
temporal objects,
temporal relations and
interactive objects. Examples of temporal objects are sounds, videos and light controls.
Temporal objects can be triggered by interactive objects (usually launched by the user) and
several temporal objects can be executed simultaneously. A temporal object may contain
other temporal objects: this hierarchy allows us to control the start or end of a temporal
object by controlling the start or end of its parent. Hierarchy is ever-present in all kinds
of music: Music pieces are often hierarchized by movements, parts, motives, measures,
among other segmentations. <ref>Mauricio Toro, Myriam Desainte-Catherine, Camilo Rueda. Formal semantics for interactive music scores: a framework to design, specify properties and execute interactive scenarios. Journal of Mathematics and Music 8 (1) </ref> <ref>{{cite web|title=Open Software System for Interactive Applications|url=https://ossia.io/|accessdate=23 January 2018|language=en-EN}}</ref>
===Orb Composer===
A program developed by Hexachords and directed by Richard Portelli, mainly focused on orchestral music.<ref>{{cite news|last1=Jallet|first1=Fabrice|title=Hexachords a créé une IA pour les compositeurs de musique / Actualités / Irma : centre d'information et de ressources pour les musiques actuelles|url=http://www.irma.asso.fr/Hexachords-a-cree-une-IA-pour-les|accessdate=23 January 2017|work=Hexachords a créé une IA pour les compositeurs de musique / Actualités / Irma : centre d'information et de ressources pour les musiques actuelles|date=13 January 2017|language=fr-FR}}</ref><ref>{{cite web|last1=Richard|first1=Portelli|title=Orb Composer Software - Music Artificial Intelligence Engine|url=http://www.hexachords.com/orb-composer|website=Hexachords|accessdate=23 January 2017|language=fr-FR}}</ref>
===EMI===
A program developed by David Cope which composes classical music. See [[Experiments in Musical Intelligence]]. [[Emily Howell]] is an interactive augmentation of EMI. (As a popular example, the background music of the [[viral video]] [[Humans Need Not Apply]] was created by "her", as revealed in the video to illustrate the likely fate of creative jobs.)
===OrchExtra===
This program was designed to provide small-budget productions with instrumentation for all instruments usually present in the full-fledged orchestra. If there is a small orchestra playing, the program can play the part for missing instruments. High school and community theaters wanting to produce a musical can now benefit from the virtual orchestra and realize a full Broadway score. This software is able to follow the fluctuations in tempo and musical expression. Musicians enjoy the thrill of playing with a full orchestra, while the audience enjoys the rich sound that comes from the combination of the virtual orchestra with the musicians.<ref>[http://www.rms.biz/index.cfm?page=OrchEXTRA Realtime Music Solutions] {{webarchive|url=https://archive.is/20060910191858/http://www.rms.biz/index.cfm?page=OrchEXTRA |date=2006-09-10 }}. Rms.biz. Retrieved on 2010-12-22.</ref><ref>[http://www.aaai.org/AITopics/pmwiki/pmwiki.php/AITopics/Music AITopics / Music]. Aaai.org. Retrieved on 2010-12-22.</ref>
Demo:<ref>[https://web.archive.org/web/20080819235351/http://www.rms.biz/Documents/Movies/OnlineDemo.htm Realtime Music Solutions]. Rms.biz. Retrieved on 2010-12-22.</ref>
===Computer Accompaniment (Carnegie Mellon University)===
The Computer Music Project at CMU develops computer music and interactive performance technology to enhance human musical experience and creativity. This interdisciplinary effort draws on [[Music Theory]], [[Cognitive Science]], [[Artificial Intelligence]] and [[Machine Learning]], [[Human Computer Interaction]], [[Real-Time Systems]], [[Computer Graphics]] and Animation, [[Multimedia]], [[Programming Languages]], and [[Signal Processing]].<ref>[http://www-2.cs.cmu.edu/~music/ Computer Music Group]. 2.cs.cmu.edu. Retrieved on 2010-12-22.</ref> One of their project is similar to SmartMusic. It provides accompaniment for the chosen piece follows the soloist (user) despite tempo changes and/or mistakes.
Demo:<ref>[http://www-2.cs.cmu.edu/~music/examples/RBDACCMP.MOV Video demo of computer accompaniment]</ref>
===SmartMusic ===
SmartMusic is an interactive, computer-based practice tool for musicians. It offers exercises, instant feedback tools, and accompaniments meant to aid musicians. The product is targeted at teachers and students alike and offers five categories of accompaniments: solo, skill development, method books, jazz, and ensemble. Teachers can give students pre-defined assignments via email and scan in sheet music that is not yet in the SmartMusic catalog. Students can choose the difficulty level they want to play at, slow down or speed up the tempo, or change the key in which to play the piece.<ref>{{Cite web|url=http://www.smartmusic.com/| title=Discover SmartMusic|website=smartmusic.com|publisher = MakeMusic, Inc.|accessdate=September 12, 2015}}</ref> SmartMusic also compares students' playing with digital template, which allows it to detect mistakes and mark them on a score. It also simulates the rapport between musicians by sensing and reacting to tempo changes.
===StarPlayIt===
StarPlay is also a music education software that allows the user to practice by performing with professional musicians, bands and orchestras. They can choose their spot and watch the video from that spot. They can hear the other musicians playing. Again, the program listens to the user's performance and helps them improve their performance by providing constructive feedback as they rehearse. StarPlay was developed by StarPlayIt (formerly In The Chair), a music technology company that has won many awards for its platforms for online musical performance and participation.<ref>[http://www.inthechair.com/index.php (BROKEN LINK) StarPlayit music learning and video game technology platform]. Inthechair.com. Retrieved on 2010-12-22.</ref>
===ChucK===
{{main|ChucK}}
Developed at Princeton University by Ge Wang and Perry Cook, ChucK is a text-based, cross-platform language that allows real-time synthesis, composition, performance and analysis of music.
.<ref>[http://chuck.cs.princeton.edu/ ChucK => Strongly-timed, On-the-fly Audio Programming Language]. Chuck.cs.princeton.edu. Retrieved on 2010-12-22.</ref> It is used by [[SLOrk]] (Stanford Laptop Orchestra) <ref>Driver, Dustin. (1999-03-26) [https://www.apple.com/pro/profiles/slork/ Pro - Profiles - Stanford Laptop Orchestra (SLOrk), pg. 1]. Apple. Retrieved on 2010-12-22.</ref> and [[PLOrk]] (Princeton Laptop Orchestra).
Demo:<ref>[https://www.apple.com/pro/profiles/slork/video.html Pro - Profiles - Stanford Laptop Orchestra (SLOrk) - Video]. Apple (1999-03-26). Retrieved on 2010-12-22.</ref>
===Impromptu===
The Impromptu<ref>[http://impromptu.moso.com.au/ impromptu]. Impromptu.moso.com.au. Retrieved on 2010-12-22.</ref> media programming environment was developed by Andrew Sorensen for exploring 'intelligent' interactive music and visual systems. Impromptu is used for live coding performances and research including generative orchestral music<ref>[http://eprints.qut.edu.au/14095/ A Computational Model For The Generation Of Orchestral Music In The Germanic Symphonic Tradition: A progress report | QUT ePrints]. Eprints.qut.edu.au. Retrieved on 2010-12-22.</ref> and computational models of music perception.<ref>[http://eprints.qut.edu.au/34417/ Interrogating statistical models of music perception | QUT ePrints]. Eprints.qut.edu.au. Retrieved on 2010-12-22.</ref>
===REAPER's TabEditor===
MIDI to string instrument (guitar, violin, dombra, etc.) tablature conversion is a nontrivial task, as the same note can reside on different strings of the instrument. And the creation of good fingering is sometimes a challenge even for real musicians, especially when translating a two handed piano composition on a string instrument.
So in TabEditor<ref>[http://wiki.cockos.com/wiki/index.php/TabEditor TabEditor]</ref> (the tiny plugin for [[REAPER]] DAW), an AI was used that solves this puzzle the same way as a musician would: trying to keep all the notes close to each other (to be possible to play) while trying to fit all the piano notes into a range that can be played simultaneously on the instrument. When direct translation is impossible (piano part has more notes than are possible on the guitar) the AI tries to find an acceptable solution, removing as few notes as possible from the original composition.
The Prolog programming language was used to create this AI.
===Ludwig===
Ludwig<ref>[http://www.write-music.com The automatic composition software Ludwig]. www.write-music.com. Retrieved on 2011-10-27.</ref> is an automated composition software based on tree search algorithms. Ludwig generates melodies according to principles of classical music theory. The software arranges its melodies with pop-automation patterns or in four-part choral writing. Ludwig can react in real-time on an eight-bar theme played on a keyboard. The theme will be analysed for key, harmonic content and rhythm while it is being performed by a human. The program then without delay repeats the theme arranged e.g. for orchestra. It subsequently varies the melody to create a little piece as interactive answer to the human input.
===OMax===
OMax<ref>[http://omax.ircam.fr/ The OMax Project Page]</ref> is a software environment which learns in real-time typical features of a musician's style and plays along with him interactively, giving the flavor of a machine co-improvisation. OMax uses OpenMusic and Max. It is based on researches on stylistic modeling carried out by Gerard Assayag and Shlomo Dubnov and on researches on improvisation with the computer by G. Assayag, M. Chemillier and G. Bloch (Aka the OMax Brothers) in the [[Ircam]] Music Representations group.
===Melomics===
Melomics is a proprietary computational system for the automatic (without human intervention) composition of music, based on bioinspired methods and produced by Melomics Media. Composing a wide variety of genres, all music composed by Melomics algorithms are available in MP3, MIDI, MusicXML, and PDF (of sheet music), after purchase. Music composed by this algorithm was organized into an album named [[Iamus (album)]], which was hailed by New Scientist as "The first complete album to be composed solely by a computer and recorded by human musicians."<ref>{{cite journal | title= Computer composer honours Turing's centenary| journal= New Scientist| date =5 July 2012| url =https://www.newscientist.com/article/mg21528724.300-computer-composer-honours-turings-centenary.html}}</ref>
===MorpheuS===
MorpheuS<ref>{{cite journal| title = MorpheuS: Automatic music generation with recurrent pattern constraints and tension profiles
| author = D. Herremans, E. Chew
| year = 2016
| journal = IEEE Transactions on Affective Computing
| volume = PP(1)
| doi = 10.1109/TAFFC.2017.2737984}}</ref> is a research project by [[Dorien Herremans]] and Elaine Chew at [[Queen Mary University of London]], funded by a Marie Skłodowská-Curie EU project. The system uses an optimization approach based on a variable neighborhood search algorithm to morph existing template pieces into novel pieces with a set level of tonal tension that changes dynamically throughout the piece. This optimization approach allows for the integration of a pattern detection technique in order to enforce long term structure and recurring themes in the generated music. Pieces composed by MorpheuS have been performed at concerts in both Stanford and London.
===Flow Machines===
Flow Machines is a research project funded by the European Research Council (ERC) and led by [[François Pachet]]. Flow Machines aims at transforming musical style into a computational object to apply to AI-generated melodies and harmonies. Flow Machines has composed two fully-fledged pop songs, issued from a collaboration between the AI software and pop composer Benoît Carré: Daddy's Car <ref>[https://www.youtube.com/watch?v=LSHZ_b05W7o]. Daddy's Car</ref> and Mister Shadow.<ref>[https://www.youtube.com/watch?v=lcGYEXJqun8]. Mister Shadow</ref>
Flow Machines also produced DeepBach, a neural network system which produces harmonisation in Bach style indiscernible from original Bach's harmonisations.<ref>[https://arxiv.org/pdf/1612.01010.pdf].</ref>
===AIVA===
Created in February 2016, in [[Luxembourg]], AIVA specializes in classical and symphonic music composition. It became the world’s first virtual composer to be recognized by a music society ([[SACEM]]). By reading a large collection of existing works of classical music (written by human composers such as Bach, Beethoven, Mozart) AIVA is capable of understanding concepts of music theory and composing on its own. The algorithm AIVA is based on deep learning and reinforcement learning architectures<ref>[http://www.aiva.ai]. AIVA 2016</ref>
===AlgoTunes===
AlgoTunes is a music software company building apps that generate music. On https://www.algotunes.com, anyone can signup to generate music of their preferred mood and style with the push of a button. The music is composed instantly in a few seconds by AlgoTunes' web app, and available for download as WAV or MIDI files.<ref>[https://www.algotunes.com]. AlgoTunes</ref>
===Amper Music===
==See also==
* [[Algorithmic composition]]
* [[Computational models of musical creativity]]
* [[Applications of artificial intelligence]]
* [[List of music software]]
==References==
{{Reflist|30em}}
==Further reading==
*[http://www.aaai.org/Press/Books/balaban.php Understanding Music with AI: Perspectives on Music Cognition]. Edited by Mira Balaban, Kemal Ebcioglu, and Otto Laske. AAAI Press.
*[http://portal.acm.org/citation.cfm?id=647303&coll=GUIDE&dl=GUIDE Proceedings of a Workshop held as part of AI-ED 93], World Conference on Artificial Intelligence in Education on Music Education: An Artificial Intelligence Approach
==External links==
*[http://www.dai.ed.ac.uk/groups/aimusic/ The Music Informatics Research Group]
*[http://www.ircam.fr/accueil.html?&L=1 Institut de Recherche et Coordination Acoustique/ Musique]
*[http://www.leeds.ac.uk/icsrim/ Interdisciplinary Centre for Research in Music]
{{DEFAULTSORT:Music And Artificial Intelligence}}
[[Category:Artificial intelligence]]
[[Category:Cognitive musicology]]
[[Category:Computer music]]' |
New page wikitext, after the edit (new_wikitext ) | '{{multiple issues|
{{cleanup|date=December 2010}}
{{advert|date=December 2010}}
}}
Research in [[artificial intelligence (AI)]] is known to have impacted [[medical diagnosis]], [[Algorithmic trading|stock trading]], [[robot control]], and several other fields. Perhaps less popular is the contribution of AI in the field of [[music]]. Nevertheless, artificial intelligence and music (AIM) has, for a long time, been a common subject in several conferences and workshops, including the [[International Computer Music Conference]], the Computing Society Conference <ref>[http://www-rcf.usc.edu/~echew/INFORMS/ics2005.html INFORMS Computing Society Conference: Annapolis: Music, Computation and AI] {{webarchive|url=https://archive.is/20120630221600/http://www-rcf.usc.edu/~echew/INFORMS/ics2005.html |date=2012-06-30 }}. Rcf.usc.edu. Retrieved on 2010-12-22.</ref> and the [[International Joint Conference on Artificial Intelligence]]. In fact, the first International Computer Music Conference was the ICMC 1974, Michigan State University, East Lansing, USA <ref>[http://www.computermusic.org/page/23/ International Computer Music Association - ICMC]. Computermusic.org (2010-11-15). Retrieved on 2010-12-22.</ref>
Current research includes the application of AI in [[Musical composition|music composition]], [[performance]], theory and digital [[Audio signal processing|sound processing]].<ref>[http://www.informaworld.com/smpp/content~db=all~content=a714908257?words=artificial*|intelligence*|music* Research in Music and Artificial Intelligence - Ergonomics Abstracts]. Informaworld.com. Retrieved on 2010-12-22.</ref>
Several music [[software]]<nowiki/>s have been developed that use AI to produce music<ref>{{cite journal| title = A Functional Taxonomy of Music Generation Systems
| author = D. Herremans, C.H., Chuan, E. Chew
| year = 2017
| journal = ACM Computing Surveys
| volume = 50
| issue = 5
| doi = 10.1109/TAFFC.2017.2737984
| pages = 69:1-30}}</ref>. Like its applications in other fields, the A.I. in this case also simulates mental task. A prominent feature is the capability of the A.I. algorithm to learn based on information obtained such as the computer accompaniment technology, which is capable of listening to and following a human performer so it can perform in synchrony.<ref>{{Cite web|url=https://pdfs.semanticscholar.org/f275/4c359d7ef052ab5997d71dc3e9443404565a.pdf|title=Artificial Intelligence, Machine Learning, and Music Understanding|last=Dannenberg|first=Roger|date=|website=Semantic Scholar|archive-url=|archive-date=|dead-url=|access-date=August 23, 2018}}</ref> Artificial intelligence also drives the so-called interactive composition technology, wherein a computer composes music in response to the performance of a live musician. There are several other A.I. applications to music that covers not only music composition, production, and performance but also the way it is marketed and consumed. Companies like [[Apple Inc.|Apple]] and [[Spotify]] rely on user data to augment their engagement metrics to power their ability to get consumers to listen more to music or push the right songs to users according to their preferences.<ref>{{Cite news|url=https://www.forbes.com/sites/kristinwestcottgrant/2017/12/10/how-to-think-about-artificial-intelligence-in-the-music-industry/#55a5c0eb7d4a|title=How To Think About Artificial Intelligence In The Music Industry|last=Grant|first=Kristin Westcott|work=Forbes|access-date=2018-08-23|language=en}}</ref>
==History==
In 1960, Russian researcher R.Kh.Zaripov published worldwide first paper on algorithmic music composing using the "[[Ural (computer)|Ural-1]]" computer.<ref>{{cite journal|last=Zaripov|first=R.Kh.|title=Об алгоритмическом описании процесса сочинения музыки (On algorithmic description of process of music composition)|journal=[[Proceedings of the USSR Academy of Sciences]]|year=1960|volume=132|issue=6}}</ref>
In 1965, inventor Ray Kurzweil premiered a piano piece created by a computer that was capable of pattern recognition in various compositions. The computer was then able to analyze and use these patterns to create novel melodies. The computer was debuted on Steve Allen's I've Got a Secret program, and stumped the hosts until film star Henry Morgan guessed Ray's secret.<ref>http://www.kurzweiltech.com/raybio.html</ref>
==Software applications==
===Interactive scores===
Multimedia Scenarios in interactive scores are represented by
temporal objects,
temporal relations and
interactive objects. Examples of temporal objects are sounds, videos and light controls.
Temporal objects can be triggered by interactive objects (usually launched by the user) and
several temporal objects can be executed simultaneously. A temporal object may contain
other temporal objects: this hierarchy allows us to control the start or end of a temporal
object by controlling the start or end of its parent. Hierarchy is ever-present in all kinds
of music: Music pieces are often hierarchized by movements, parts, motives, measures,
among other segmentations. <ref>Mauricio Toro, Myriam Desainte-Catherine, Camilo Rueda. Formal semantics for interactive music scores: a framework to design, specify properties and execute interactive scenarios. Journal of Mathematics and Music 8 (1) </ref> <ref>{{cite web|title=Open Software System for Interactive Applications|url=https://ossia.io/|accessdate=23 January 2018|language=en-EN}}</ref>
===Orb Composer===
A program developed by Hexachords and directed by Richard Portelli, mainly focused on orchestral music.<ref>{{cite news|last1=Jallet|first1=Fabrice|title=Hexachords a créé une IA pour les compositeurs de musique / Actualités / Irma : centre d'information et de ressources pour les musiques actuelles|url=http://www.irma.asso.fr/Hexachords-a-cree-une-IA-pour-les|accessdate=23 January 2017|work=Hexachords a créé une IA pour les compositeurs de musique / Actualités / Irma : centre d'information et de ressources pour les musiques actuelles|date=13 January 2017|language=fr-FR}}</ref><ref>{{cite web|last1=Richard|first1=Portelli|title=Orb Composer Software - Music Artificial Intelligence Engine|url=http://www.hexachords.com/orb-composer|website=Hexachords|accessdate=23 January 2017|language=fr-FR}}</ref>
===EMI===
A program developed by David Cope which composes classical music. See [[Experiments in Musical Intelligence]]. [[Emily Howell]] is an interactive augmentation of EMI. (As a popular example, the background music of the [[viral video]] [[Humans Need Not Apply]] was created by "her", as revealed in the video to illustrate the likely fate of creative jobs.)
===OrchExtra===
This program was designed to provide small-budget productions with instrumentation for all instruments usually present in the full-fledged orchestra. If there is a small orchestra playing, the program can play the part for missing instruments. High school and community theaters wanting to produce a musical can now benefit from the virtual orchestra and realize a full Broadway score. This software is able to follow the fluctuations in tempo and musical expression. Musicians enjoy the thrill of playing with a full orchestra, while the audience enjoys the rich sound that comes from the combination of the virtual orchestra with the musicians.<ref>[http://www.rms.biz/index.cfm?page=OrchEXTRA Realtime Music Solutions] {{webarchive|url=https://archive.is/20060910191858/http://www.rms.biz/index.cfm?page=OrchEXTRA |date=2006-09-10 }}. Rms.biz. Retrieved on 2010-12-22.</ref><ref>[http://www.aaai.org/AITopics/pmwiki/pmwiki.php/AITopics/Music AITopics / Music]. Aaai.org. Retrieved on 2010-12-22.</ref>
Demo:<ref>[https://web.archive.org/web/20080819235351/http://www.rms.biz/Documents/Movies/OnlineDemo.htm Realtime Music Solutions]. Rms.biz. Retrieved on 2010-12-22.</ref>
===Computer Accompaniment (Carnegie Mellon University)===
The Computer Music Project at CMU develops computer music and interactive performance technology to enhance human musical experience and creativity. This interdisciplinary effort draws on [[Music Theory]], [[Cognitive Science]], [[Artificial Intelligence]] and [[Machine Learning]], [[Human Computer Interaction]], [[Real-Time Systems]], [[Computer Graphics]] and Animation, [[Multimedia]], [[Programming Languages]], and [[Signal Processing]].<ref>[http://www-2.cs.cmu.edu/~music/ Computer Music Group]. 2.cs.cmu.edu. Retrieved on 2010-12-22.</ref> One of their project is similar to SmartMusic. It provides accompaniment for the chosen piece follows the soloist (user) despite tempo changes and/or mistakes.
Demo:<ref>[http://www-2.cs.cmu.edu/~music/examples/RBDACCMP.MOV Video demo of computer accompaniment]</ref>
===SmartMusic ===
SmartMusic is an interactive, computer-based practice tool for musicians. It offers exercises, instant feedback tools, and accompaniments meant to aid musicians. The product is targeted at teachers and students alike and offers five categories of accompaniments: solo, skill development, method books, jazz, and ensemble. Teachers can give students pre-defined assignments via email and scan in sheet music that is not yet in the SmartMusic catalog. Students can choose the difficulty level they want to play at, slow down or speed up the tempo, or change the key in which to play the piece.<ref>{{Cite web|url=http://www.smartmusic.com/| title=Discover SmartMusic|website=smartmusic.com|publisher = MakeMusic, Inc.|accessdate=September 12, 2015}}</ref> SmartMusic also compares students' playing with digital template, which allows it to detect mistakes and mark them on a score. It also simulates the rapport between musicians by sensing and reacting to tempo changes.
===StarPlayIt===
StarPlay is also a music education software that allows the user to practice by performing with professional musicians, bands and orchestras. They can choose their spot and watch the video from that spot. They can hear the other musicians playing. Again, the program listens to the user's performance and helps them improve their performance by providing constructive feedback as they rehearse. StarPlay was developed by StarPlayIt (formerly In The Chair), a music technology company that has won many awards for its platforms for online musical performance and participation.<ref>[http://www.inthechair.com/index.php (BROKEN LINK) StarPlayit music learning and video game technology platform]. Inthechair.com. Retrieved on 2010-12-22.</ref>
===ChucK===
{{main|ChucK}}
Developed at Princeton University by Ge Wang and Perry Cook, ChucK is a text-based, cross-platform language that allows real-time synthesis, composition, performance and analysis of music.
.<ref>[http://chuck.cs.princeton.edu/ ChucK => Strongly-timed, On-the-fly Audio Programming Language]. Chuck.cs.princeton.edu. Retrieved on 2010-12-22.</ref> It is used by [[SLOrk]] (Stanford Laptop Orchestra) <ref>Driver, Dustin. (1999-03-26) [https://www.apple.com/pro/profiles/slork/ Pro - Profiles - Stanford Laptop Orchestra (SLOrk), pg. 1]. Apple. Retrieved on 2010-12-22.</ref> and [[PLOrk]] (Princeton Laptop Orchestra).
Demo:<ref>[https://www.apple.com/pro/profiles/slork/video.html Pro - Profiles - Stanford Laptop Orchestra (SLOrk) - Video]. Apple (1999-03-26). Retrieved on 2010-12-22.</ref>
===Impromptu===
The Impromptu<ref>[http://impromptu.moso.com.au/ impromptu]. Impromptu.moso.com.au. Retrieved on 2010-12-22.</ref> media programming environment was developed by Andrew Sorensen for exploring 'intelligent' interactive music and visual systems. Impromptu is used for live coding performances and research including generative orchestral music<ref>[http://eprints.qut.edu.au/14095/ A Computational Model For The Generation Of Orchestral Music In The Germanic Symphonic Tradition: A progress report | QUT ePrints]. Eprints.qut.edu.au. Retrieved on 2010-12-22.</ref> and computational models of music perception.<ref>[http://eprints.qut.edu.au/34417/ Interrogating statistical models of music perception | QUT ePrints]. Eprints.qut.edu.au. Retrieved on 2010-12-22.</ref>
===REAPER's TabEditor===
MIDI to string instrument (guitar, violin, dombra, etc.) tablature conversion is a nontrivial task, as the same note can reside on different strings of the instrument. And the creation of good fingering is sometimes a challenge even for real musicians, especially when translating a two handed piano composition on a string instrument.
So in TabEditor<ref>[http://wiki.cockos.com/wiki/index.php/TabEditor TabEditor]</ref> (the tiny plugin for [[REAPER]] DAW), an AI was used that solves this puzzle the same way as a musician would: trying to keep all the notes close to each other (to be possible to play) while trying to fit all the piano notes into a range that can be played simultaneously on the instrument. When direct translation is impossible (piano part has more notes than are possible on the guitar) the AI tries to find an acceptable solution, removing as few notes as possible from the original composition.
The Prolog programming language was used to create this AI.
===Ludwig===
Ludwig<ref>[http://www.write-music.com The automatic composition software Ludwig]. www.write-music.com. Retrieved on 2011-10-27.</ref> is an automated composition software based on tree search algorithms. Ludwig generates melodies according to principles of classical music theory. The software arranges its melodies with pop-automation patterns or in four-part choral writing. Ludwig can react in real-time on an eight-bar theme played on a keyboard. The theme will be analysed for key, harmonic content and rhythm while it is being performed by a human. The program then without delay repeats the theme arranged e.g. for orchestra. It subsequently varies the melody to create a little piece as interactive answer to the human input.
===OMax===
OMax<ref>[http://omax.ircam.fr/ The OMax Project Page]</ref> is a software environment which learns in real-time typical features of a musician's style and plays along with him interactively, giving the flavor of a machine co-improvisation. OMax uses OpenMusic and Max. It is based on researches on stylistic modeling carried out by Gerard Assayag and Shlomo Dubnov and on researches on improvisation with the computer by G. Assayag, M. Chemillier and G. Bloch (Aka the OMax Brothers) in the [[Ircam]] Music Representations group.
===Melomics===
Melomics is a proprietary computational system for the automatic (without human intervention) composition of music, based on bioinspired methods and produced by Melomics Media. Composing a wide variety of genres, all music composed by Melomics algorithms are available in MP3, MIDI, MusicXML, and PDF (of sheet music), after purchase. Music composed by this algorithm was organized into an album named [[Iamus (album)]], which was hailed by New Scientist as "The first complete album to be composed solely by a computer and recorded by human musicians."<ref>{{cite journal | title= Computer composer honours Turing's centenary| journal= New Scientist| date =5 July 2012| url =https://www.newscientist.com/article/mg21528724.300-computer-composer-honours-turings-centenary.html}}</ref>
===MorpheuS===
MorpheuS<ref>{{cite journal| title = MorpheuS: Automatic music generation with recurrent pattern constraints and tension profiles
| author = D. Herremans, E. Chew
| year = 2016
| journal = IEEE Transactions on Affective Computing
| volume = PP(1)
| doi = 10.1109/TAFFC.2017.2737984}}</ref> is a research project by [[Dorien Herremans]] and Elaine Chew at [[Queen Mary University of London]], funded by a Marie Skłodowská-Curie EU project. The system uses an optimization approach based on a variable neighborhood search algorithm to morph existing template pieces into novel pieces with a set level of tonal tension that changes dynamically throughout the piece. This optimization approach allows for the integration of a pattern detection technique in order to enforce long term structure and recurring themes in the generated music. Pieces composed by MorpheuS have been performed at concerts in both Stanford and London.
===Flow Machines===
Flow Machines is a research project funded by the European Research Council (ERC) and led by [[François Pachet]]. Flow Machines aims at transforming musical style into a computational object to apply to AI-generated melodies and harmonies. Flow Machines has composed two fully-fledged pop songs, issued from a collaboration between the AI software and pop composer Benoît Carré: Daddy's Car <ref>[https://www.youtube.com/watch?v=LSHZ_b05W7o]. Daddy's Car</ref> and Mister Shadow.<ref>[https://www.youtube.com/watch?v=lcGYEXJqun8]. Mister Shadow</ref>
Flow Machines also produced DeepBach, a neural network system which produces harmonisation in Bach style indiscernible from original Bach's harmonisations.<ref>[https://arxiv.org/pdf/1612.01010.pdf].</ref>
===AIVA===
Created in February 2016, in [[Luxembourg]], AIVA specializes in classical and symphonic music composition. It became the world’s first virtual composer to be recognized by a music society ([[SACEM]]). By reading a large collection of existing works of classical music (written by human composers such as Bach, Beethoven, Mozart) AIVA is capable of understanding concepts of music theory and composing on its own. The algorithm AIVA is based on deep learning and reinforcement learning architectures<ref>[http://www.aiva.ai]. AIVA 2016</ref>
===AlgoTunes===
AlgoTunes is a music software company building apps that generate music. On https://www.algotunes.com, anyone can signup to generate music of their preferred mood and style with the push of a button. The music is composed instantly in a few seconds by AlgoTunes' web app, and available for download as WAV or MIDI files.<ref>[https://www.algotunes.com]. AlgoTunes</ref>
===Amper Music===
==See also==
* [[Algorithmic composition]]
* [[Computational models of musical creativity]]
* [[Applications of artificial intelligence]]
* [[List of music software]]
==References==
{{Reflist|30em}}
==Further reading==
*[http://www.aaai.org/Press/Books/balaban.php Understanding Music with AI: Perspectives on Music Cognition]. Edited by Mira Balaban, Kemal Ebcioglu, and Otto Laske. AAAI Press.
*[http://portal.acm.org/citation.cfm?id=647303&coll=GUIDE&dl=GUIDE Proceedings of a Workshop held as part of AI-ED 93], World Conference on Artificial Intelligence in Education on Music Education: An Artificial Intelligence Approach
==External links==
*[http://www.dai.ed.ac.uk/groups/aimusic/ The Music Informatics Research Group]
*[http://www.ircam.fr/accueil.html?&L=1 Institut de Recherche et Coordination Acoustique/ Musique]
*[http://www.leeds.ac.uk/icsrim/ Interdisciplinary Centre for Research in Music]
{{DEFAULTSORT:Music And Artificial Intelligence}}
[[Category:Artificial intelligence]]
[[Category:Cognitive musicology]]
[[Category:Computer music]]' |
Unified diff of changes made by edit (edit_diff ) | '@@ -5,5 +5,7 @@
Research in [[artificial intelligence (AI)]] is known to have impacted [[medical diagnosis]], [[Algorithmic trading|stock trading]], [[robot control]], and several other fields. Perhaps less popular is the contribution of AI in the field of [[music]]. Nevertheless, artificial intelligence and music (AIM) has, for a long time, been a common subject in several conferences and workshops, including the [[International Computer Music Conference]], the Computing Society Conference <ref>[http://www-rcf.usc.edu/~echew/INFORMS/ics2005.html INFORMS Computing Society Conference: Annapolis: Music, Computation and AI] {{webarchive|url=https://archive.is/20120630221600/http://www-rcf.usc.edu/~echew/INFORMS/ics2005.html |date=2012-06-30 }}. Rcf.usc.edu. Retrieved on 2010-12-22.</ref> and the [[International Joint Conference on Artificial Intelligence]]. In fact, the first International Computer Music Conference was the ICMC 1974, Michigan State University, East Lansing, USA <ref>[http://www.computermusic.org/page/23/ International Computer Music Association - ICMC]. Computermusic.org (2010-11-15). Retrieved on 2010-12-22.</ref>
-Current research includes the application of AI in music composition, performance, theory and digital sound processing.<ref>[http://www.informaworld.com/smpp/content~db=all~content=a714908257?words=artificial*|intelligence*|music* Research in Music and Artificial Intelligence - Ergonomics Abstracts]. Informaworld.com. Retrieved on 2010-12-22.</ref> Several music software applications have been developed that use AI to produce music<ref>{{cite journal| title = A Functional Taxonomy of Music Generation Systems
+Current research includes the application of AI in [[Musical composition|music composition]], [[performance]], theory and digital [[Audio signal processing|sound processing]].<ref>[http://www.informaworld.com/smpp/content~db=all~content=a714908257?words=artificial*|intelligence*|music* Research in Music and Artificial Intelligence - Ergonomics Abstracts]. Informaworld.com. Retrieved on 2010-12-22.</ref>
+
+Several music [[software]]<nowiki/>s have been developed that use AI to produce music<ref>{{cite journal| title = A Functional Taxonomy of Music Generation Systems
| author = D. Herremans, C.H., Chuan, E. Chew
| year = 2017
@@ -12,5 +14,5 @@
| issue = 5
| doi = 10.1109/TAFFC.2017.2737984
-| pages = 69:1-30}}</ref>. A few examples are included below. Note that there are many that are still being developed.
+| pages = 69:1-30}}</ref>. Like its applications in other fields, the A.I. in this case also simulates mental task. A prominent feature is the capability of the A.I. algorithm to learn based on information obtained such as the computer accompaniment technology, which is capable of listening to and following a human performer so it can perform in synchrony.<ref>{{Cite web|url=https://pdfs.semanticscholar.org/f275/4c359d7ef052ab5997d71dc3e9443404565a.pdf|title=Artificial Intelligence, Machine Learning, and Music Understanding|last=Dannenberg|first=Roger|date=|website=Semantic Scholar|archive-url=|archive-date=|dead-url=|access-date=August 23, 2018}}</ref> Artificial intelligence also drives the so-called interactive composition technology, wherein a computer composes music in response to the performance of a live musician. There are several other A.I. applications to music that covers not only music composition, production, and performance but also the way it is marketed and consumed. Companies like [[Apple Inc.|Apple]] and [[Spotify]] rely on user data to augment their engagement metrics to power their ability to get consumers to listen more to music or push the right songs to users according to their preferences.<ref>{{Cite news|url=https://www.forbes.com/sites/kristinwestcottgrant/2017/12/10/how-to-think-about-artificial-intelligence-in-the-music-industry/#55a5c0eb7d4a|title=How To Think About Artificial Intelligence In The Music Industry|last=Grant|first=Kristin Westcott|work=Forbes|access-date=2018-08-23|language=en}}</ref>
==History==
' |
New page size (new_size ) | 18977 |
Old page size (old_size ) | 17483 |
Size change in edit (edit_delta ) | 1494 |
Lines added in edit (added_lines ) | [
0 => 'Current research includes the application of AI in [[Musical composition|music composition]], [[performance]], theory and digital [[Audio signal processing|sound processing]].<ref>[http://www.informaworld.com/smpp/content~db=all~content=a714908257?words=artificial*|intelligence*|music* Research in Music and Artificial Intelligence - Ergonomics Abstracts]. Informaworld.com. Retrieved on 2010-12-22.</ref> ',
1 => false,
2 => 'Several music [[software]]<nowiki/>s have been developed that use AI to produce music<ref>{{cite journal| title = A Functional Taxonomy of Music Generation Systems',
3 => '| pages = 69:1-30}}</ref>. Like its applications in other fields, the A.I. in this case also simulates mental task. A prominent feature is the capability of the A.I. algorithm to learn based on information obtained such as the computer accompaniment technology, which is capable of listening to and following a human performer so it can perform in synchrony.<ref>{{Cite web|url=https://pdfs.semanticscholar.org/f275/4c359d7ef052ab5997d71dc3e9443404565a.pdf|title=Artificial Intelligence, Machine Learning, and Music Understanding|last=Dannenberg|first=Roger|date=|website=Semantic Scholar|archive-url=|archive-date=|dead-url=|access-date=August 23, 2018}}</ref> Artificial intelligence also drives the so-called interactive composition technology, wherein a computer composes music in response to the performance of a live musician. There are several other A.I. applications to music that covers not only music composition, production, and performance but also the way it is marketed and consumed. Companies like [[Apple Inc.|Apple]] and [[Spotify]] rely on user data to augment their engagement metrics to power their ability to get consumers to listen more to music or push the right songs to users according to their preferences.<ref>{{Cite news|url=https://www.forbes.com/sites/kristinwestcottgrant/2017/12/10/how-to-think-about-artificial-intelligence-in-the-music-industry/#55a5c0eb7d4a|title=How To Think About Artificial Intelligence In The Music Industry|last=Grant|first=Kristin Westcott|work=Forbes|access-date=2018-08-23|language=en}}</ref> '
] |
Lines removed in edit (removed_lines ) | [
0 => 'Current research includes the application of AI in music composition, performance, theory and digital sound processing.<ref>[http://www.informaworld.com/smpp/content~db=all~content=a714908257?words=artificial*|intelligence*|music* Research in Music and Artificial Intelligence - Ergonomics Abstracts]. Informaworld.com. Retrieved on 2010-12-22.</ref> Several music software applications have been developed that use AI to produce music<ref>{{cite journal| title = A Functional Taxonomy of Music Generation Systems',
1 => '| pages = 69:1-30}}</ref>. A few examples are included below. Note that there are many that are still being developed.'
] |
Whether or not the change was made through a Tor exit node (tor_exit_node ) | false |
Unix timestamp of change (timestamp ) | 1535012213 |