Chapter 5: Tone Transfer with FM Synthesis
5.1: Differentiable FM Synthesis
DDX7 Audio Examples
Add here the DDX7 audio examples from the paper. WIP: refer to this section in the Thesis
Parameter Intervention
Add audio examples for DDX7 parameter intervention
DDX7 Tone Transfer
Playing a model trained on violin audio!
5.2 Direct Envelope Learning
Envelope Learning Audio Examples?
Show that the reconstructed audio samples sound the same?? (not refereced by thesis)
Envelope Learning Demo
Chapter 6: Representation Learning for Low-Latency Interaction
Papers and info:
Link to Paper’s Website
Link to Audio Plugin
Timbre Transfer with BRAVE
We present the reconstruction results of percussive and tonal sounds of RAVE and BRAVE models trained on the Filosax and Drumset datasets.
Models Trained on the Drumset Dataset
We encode and decode original excerpts of Drumset, Beatbox and Candombe datasets, using RAVE and BRAVE models trained on Drumset, and present the results on the following table.
| Instrument | Original | RAVE (Trained on Drumset) | BRAVE (Trained on Drumset) |
|---|---|---|---|
| Drumset | |||
| Beatbox | |||
| Candombe |
Models Trained on the Filosax Dataset
We encode and decode original excerpts of the Filosax, Svoice and Viola datasets, using RAVE and BRAVE models trained on Filosax, and present the results on the following table.
| Instrument | Original | RAVE (Trained on Filosax) | BRAVE (Trained on Filosax) |
|---|---|---|---|
| Filosax | |||
| Svoice | |||
| Viola |
Chapter 7: Interaction Design with Neural Audio
Papers and info:
Rendered Transformations for different BRAVE Configurations
| Sound Example | Acoustic Guitar: Chord and Arpeggio | Electric Guitar: Muted Strings |
|---|---|---|
| Original (Input) |
|
|
| FM Tone Transfer (BRASS) |
|
|
| BRAVE Original |
|
|
| BRAVE Two-Phase |
|
|
| BRAVE One-Phase |
|
|
| BRAVE Grafted (VCTK) |
|
|
Tonal and Percussive Rendering for Different Representations
The following table presents audio examples of transformations of different violin articulation excerpts models trained on the same data, but which learn different representations.
It seems that one-phase models, and models grafted with one-phase encoders, learn representations with a good capacity of tonal rendering, where a tonal input seems to “bleed” through the model (highlighted in blue). Models trained on a two-phase approach, or grafted with a two-phase encoder, seem to render atonal outputs even in presence of tonal inputs (in orange), a phenomenom through which the model seems to “hallucinate” atonal sounds in response to a tonal input.
| Bowing behind the bridge | Ordinario | Jeté/Ricochet | ||
|---|---|---|---|---|
| Input | ||||
| Animals | Two-Phase | |||
| One-Phase | ||||
| Grafted (one-phase) | ||||
| Dice | Two-Phase | |||
| Grafted (two-phase) | ||||
| One-Phase | ||||
| Music | Two-Phase | |||
| One-Phase |