Very interesting. I recently found out the house I grew up in in the UK over 15+ years had lead pipes that were never replaced, and I always wondered why they were not replaced and if it had some cognitive impact. This lining effect likely explains the reason, and offers at least some reassurance.
Have you tried the Yoto player? Our daughter loves hers and they are pretty cheap. You can load them with your own mp3s too, which we download from BBC Sounds.
My favourite part of this is when they apply their new words to things that technically make sense, but don't. My daughter proudly pointed at a king wearing a crown as "sharp king" after learning about knives, saws, etc.
One side effect of this I remember was that, for a little while, there was a blackmarket of buying MP3s on CD at car boots and flea markets. My dad came home once with a CD with all the Beatles songs. He had paid something like £15 and thought he'd got a deal of his life.
I also wonder which year it stopped being an acceptable Christmas present to give someone a burned CD.
Awesome work! One thing I'm curious about in this space is why people generally generate the sound form directly. I always imagined you'd get better results teaching the model to output parameters which you could feed into synths (wavetable/fm/granular/VA), samplers, and effects, alongside MIDI.
You'd imagine you could estimate most music with this with less compute and higher determinism and introspection. Is it because there isn't enough training data for the above?
Thank you! There has been a lot of work on midi generation in the past and many people have gotten great results using the approach that you describe. The reason why modern music generators like ours create audio files directly is because midi can't represent all of the nuances of acoustic instruments (vocals, violin, trombone, etc). The allure of modern generative AI (diffusion networks and autoregressive models) is that they are finally capable of generating high quality audio which sounds natural.
If you're interested in really exciting work on applying AI to creating synthesizer patches, I recommend you check out synplant2: https://soniccharge.com/synplant2. Their tool can load in any audio and then create a synth patch which sounds nearly identical to the input audio.
That's a solid point about the limitations of MIDI, @kantthpel. Synplant2 sounds like a neat bridge between traditional synthesis and AI's capabilities. Wonder if it could lead to a hybrid approach where AI-generated parameters enhance MIDI compositions, making them sound more natural without fully ditching the efficiency of MIDI. Could be a game-changer for composers looking to blend electronic and acoustic sounds.
reply