In music creation, rapid prototyping is essential for exploring and refining ideas, yet existing generative tools often fall short when users require both structural control and stylistic flexibility. Prior approaches in stem-to-stem generation can condition on other musical stems but offer limited control over rhythm, and timbre-transfer methods allow users to specify specific rhythms, but cannot condition on musical context. We introduce DARC, a generative drum accompaniment model that conditions both on musical context from other stems and explicit rhythm prompts such as beatboxing or tapping tracks. Using parameter-efficient fine-tuning, we augment STAGE, a state-of-the-art drum stem generator, with fine-grained rhythm control while maintaining musical context awareness.
翻译:在音乐创作中,快速原型制作对于探索和完善创意至关重要,然而现有的生成工具在用户同时需要结构控制和风格灵活性时往往表现不足。先前的音轨到音轨生成方法能够以其他音乐音轨为条件,但对节奏的控制有限;音色转换方法允许用户指定特定节奏,却无法以音乐上下文为条件。本文提出DARC,一种生成式鼓伴奏模型,它同时以其他音轨的音乐上下文和明确的节奏提示(如节奏口技或敲击音轨)为条件。通过参数高效微调,我们在保持音乐上下文感知能力的同时,为最先进的鼓音轨生成模型STAGE增强了细粒度节奏控制功能。