Really elegant approach - mapping sentence embeddings to a deterministic synth feels more like building an instrument than generating content, and the instant playback makes it great for flow.
Would love to know if the same prompt always yields the same sound (reproducibility could be powerful), and whether you’ve considered semantic morphing between two moods over time.
Thanks!
The same prompt yields largely the same song because the 'Fast' (default) mode retrieves the synth parameters from a pre-existing library.
But if you use the 'Custom LLM' model, it can generate new and creative music every time you play something - even for the same input!