NSynth: Difference between revisions
Rustycandle (talk | contribs) No edit summary |
Rustycandle (talk | contribs) No edit summary |
||
Line 13: | Line 13: | ||
}} |
}} |
||
'''NSynth''' (Neural Synthesizer) is |
'''NSynth''' (a [[portmanteau]] of "Neural Synthesizer") is a software [[algorithm]], outlined in a paper in April 2017<ref name="auto">{{Cite arXiv|eprint=1704.01279|class=cs.LG|first1=Jesse|last1=Engel|first2=Cinjon|last2=Resnick|title=Neural Audio Synthesis of Musical Notes with WaveNet Autoencoders|last3=Roberts|first3=Adam|last4=Dieleman|first4=Sander|first5=Douglas|last6=Simonyan|first6=Karen|last7=Norouzi|first7=Mohammad|year=2017|last5=Eck}}</ref>, that generates new sounds through a [[neural network]] based [[synthesis]], employing a [[WaveNet]]-style [[autoencoder]] to learn its own temporal embeddings from four different sounds.<ref>{{Cite web|title=Neural Audio Synthesis of Musical Notes with WaveNet Autoencoders|url=https://research.google/pubs/pub46119/|website=research.google}}</ref>. Google then released an open source hardware interface for the algorithm called NSynth Super<ref>{{Cite web|title=Google's open-source neural synth is creating totally new sounds|url=https://www.wired.co.uk/article/google-ai-nsynth-algorithm-music-creativity|website=wired.co.uk}}</ref>, used by musicians such as [[Grimes]]<ref>{{Cite web|title=73 | Grimes (c) on Music, Creativity, and Digital Personae – Sean Carroll|url=https://www.preposterousuniverse.com/podcast/2019/11/18/73-grimes-c-on-music-creativity-and-digital-personae/|website=www.preposterousuniverse.com}}</ref> and [[Yacht (band)|YACHT]] <ref>{{Cite web|title= Music and Machine Learning (Google I/O'19) |url=https://www.youtube.com/watch?v=pM9u9xcM_cs|website=youtube.com}}</ref> to generate experimental music using AI. The research and development of the algorithm was part of a collaboration between [[Google Brain]], [[Magenta (research team) | Magenta]] and [[DeepMind]].<ref>{{Cite web|title=NSynth: Neural Audio Synthesis|url=https://magenta.tensorflow.org/nsynth|website=Magenta}}</ref> |
||
== |
== Development == |
||
=== Dataset === |
=== Dataset === |
||
Line 22: | Line 22: | ||
=== Model === |
=== Model === |
||
A spectral autoencoder model and a WaveNet autoencoder model are publicly available on GitHub <ref>{{Cite web|title=NSynth: Neural Audio Synthesis|url= |
|||
https://github.com/magenta/magenta/tree/main/magenta/models/nsynth|website=GitHub}}</ref>. The baseline model uses a spectrogram with fft_size 1024 and hop_size 256, MSE loss on the magnitudes, and the Griffin-Lim algorithm for reconstruction. The WaveNet model trains on mu-law encoded waveform chunks of size 6144. It learns embeddings with 16 dimensions that are downsampled by 512 in time<ref name="auto">{{Cite arXiv|eprint=1704.01279|class=cs.LG|first1=Jesse|last1=Engel|first2=Cinjon|last2=Resnick|title=Neural Audio Synthesis of Musical Notes with WaveNet Autoencoders|last3=Roberts|first3=Adam|last4=Dieleman|first4=Sander|first5=Douglas|last6=Simonyan|first6=Karen|last7=Norouzi|first7=Mohammad|year=2017|last5=Eck}}</ref>. |
|||
{{Infobox synthesizer |
{{Infobox synthesizer |
Revision as of 09:57, 2 November 2022
This article, NSynth, has recently been created via the Articles for creation process. Please check to see if the reviewer has accidentally left this template after accepting the draft and take appropriate action as necessary.
Reviewer tools: Inform author |
Original author(s) | Google Brain, Deep Mind, Magenta |
---|---|
Initial release | 6 April 2017 |
Repository | github |
Written in | Python |
Type | Software synthesizer |
License | Apache 2.0 |
Website | magenta |
NSynth (a portmanteau of "Neural Synthesizer") is a software algorithm, outlined in a paper in April 2017[1], that generates new sounds through a neural network based synthesis, employing a WaveNet-style autoencoder to learn its own temporal embeddings from four different sounds.[2]. Google then released an open source hardware interface for the algorithm called NSynth Super[3], used by musicians such as Grimes[4] and YACHT [5] to generate experimental music using AI. The research and development of the algorithm was part of a collaboration between Google Brain, Magenta and DeepMind.[6]
Development
Dataset
The NSynth dataset is composed of 305,979 one-shot instrumental notes featuring a unique pitch, timbre, and envelope, sampled from 1,006 instruments from commercial sample libraries.[7] For each instrument the dataset contains four-second 16kHz audio snippets by ranging over every pitch of a standard MIDI piano, as well as five different velocities [8]. The dataset is made available under a Creative Commons Attribution 4.0 International (CC BY 4.0) license. [9]
Model
A spectral autoencoder model and a WaveNet autoencoder model are publicly available on GitHub [10]. The baseline model uses a spectrogram with fft_size 1024 and hop_size 256, MSE loss on the magnitudes, and the Griffin-Lim algorithm for reconstruction. The WaveNet model trains on mu-law encoded waveform chunks of size 6144. It learns embeddings with 16 dimensions that are downsampled by 512 in time[1].
NSynth Super | |
---|---|
Manufacturer | Google Brain, Google Creative Lab |
Dates | 2018 |
Technical specifications | |
Synthesis type | Neural Network Sample-based synthesis |
Input/output | |
Left-hand control | Pitch bend, ADSR |
External control | MIDI |
Nsynth Super
Nsynth Super is an audio synthesizer released by Google in 2018.[11] It is designed to provide musicians with an accessible interface to the Nsynth algorithm.[1] Design files, source code and internal components are released under an open source Apache License 2.0[12], enabling hobbyists and musicians to freely build and use the instrument.[13][14] The instrument includes features from notable artists, such as Grimes and Yacht, using Nsynth Super in their music productions.[15]
Hardware
I'm baby four dollar toast deep v subway tile small batch affogato. Celiac bitters cray post-ironic, DSA coloring book sustainable whatever. Cronut trust fund lo-fi, ugh flexitarian dreamcatcher paleo. XOXO mumblecore listicle man braid lomo poke blog. Photo booth cornhole mukbang edison bulb, put a bird on it 3 wolf moon
Influence
google IO / Grimes etc baby four dollar toast deep v subway tile small batch affogato. Celiac bitters cray post-ironic, DSA coloring book sustainable whatever. Cronut trust fund lo-fi, ugh flexitarian dreamcatcher paleo. XOXO mumblecore listicle man braid lomo poke blog. Photo booth cornhole mukbang edison bulb, put a bird on it 3 wolf moon
References
- ^ a b c Engel, Jesse; Resnick, Cinjon; Roberts, Adam; Dieleman, Sander; Eck, Douglas; Simonyan, Karen; Norouzi, Mohammad (2017). "Neural Audio Synthesis of Musical Notes with WaveNet Autoencoders". arXiv:1704.01279 [cs.LG].
- ^ "Neural Audio Synthesis of Musical Notes with WaveNet Autoencoders". research.google.
- ^ "Google's open-source neural synth is creating totally new sounds". wired.co.uk.
- ^ "73 | Grimes (c) on Music, Creativity, and Digital Personae – Sean Carroll". www.preposterousuniverse.com.
- ^ "Music and Machine Learning (Google I/O'19)". youtube.com.
- ^ "NSynth: Neural Audio Synthesis". Magenta.
- ^ "NSynth Dataset". activeloop.ai.
- ^ A bot will complete this citation soon. Click here to jump the queue arXiv:1907.08520.
- ^ "The NSynth Dataset". tensorflow.org.
- ^ "NSynth: Neural Audio Synthesis". GitHub.
- ^ "Google built a musical instrument that uses AI and released the plans so you can make your own". CNBC. 13 March 2018.
- ^ "googlecreativelab/open-nsynth-super". April 1, 2021 – via GitHub.
- ^ Nast, Condé. "Google's open-source neural synth is creating totally new sounds". Wired UK.
- ^ "Open NSynth Super". hackaday.io.
- ^ "73 | Grimes (c) on Music, Creativity, and Digital Personae – Sean Carroll". www.preposterousuniverse.com.
Further reading
Engel, Jesse; Resnick, Cinjon; Roberts, Adam; Dieleman, Sander; Eck, Douglas; Simonyan, Karen; Norouzi, Mohammad (2017). "Neural Audio Synthesis of Musical Notes with WaveNet Autoencoders". arXiv:1704.01279 [cs.LG].
External links