mirror of
https://github.com/eliasstepanik/strudel-docker.git
synced 2026-01-11 21:58:31 +00:00
+ refactor sampler to use it + refactor synth to use it + add 'source' control + wip: samples tab + wip: webadirt ? + wip: soundfonts
225 lines
6.9 KiB
JavaScript
225 lines
6.9 KiB
JavaScript
/*
|
|
webaudio.mjs - <short description TODO>
|
|
Copyright (C) 2022 Strudel contributors - see <https://github.com/tidalcycles/strudel/blob/main/packages/webaudio/webaudio.mjs>
|
|
This program is free software: you can redistribute it and/or modify it under the terms of the GNU Affero General Public License as published by the Free Software Foundation, either version 3 of the License, or (at your option) any later version. This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Affero General Public License for more details. You should have received a copy of the GNU Affero General Public License along with this program. If not, see <https://www.gnu.org/licenses/>.
|
|
*/
|
|
|
|
import * as strudel from '@strudel.cycles/core';
|
|
import './feedbackdelay.mjs';
|
|
import './reverb.mjs';
|
|
const { Pattern } = strudel;
|
|
import './vowel.mjs';
|
|
import workletsUrl from './worklets.mjs?url';
|
|
import { getFilter, gainNode } from './helpers.mjs';
|
|
|
|
// export const getAudioContext = () => Tone.getContext().rawContext;
|
|
|
|
export const soundMap = new Map();
|
|
// onTrigger = ({ hap: Hap, t: number, deadline: number, duration: number, cps: number }) => AudioNode
|
|
export function setSound(key, onTrigger) {
|
|
soundMap.set(key, onTrigger);
|
|
}
|
|
export const resetLoadedSounds = () => soundMap.clear();
|
|
|
|
let audioContext;
|
|
export const getAudioContext = () => {
|
|
if (!audioContext) {
|
|
audioContext = new AudioContext();
|
|
}
|
|
return audioContext;
|
|
};
|
|
|
|
let destination;
|
|
const getDestination = () => {
|
|
const ctx = getAudioContext();
|
|
if (!destination) {
|
|
destination = ctx.createGain();
|
|
destination.connect(ctx.destination);
|
|
}
|
|
return destination;
|
|
};
|
|
|
|
export const panic = () => {
|
|
getDestination().gain.linearRampToValueAtTime(0, getAudioContext().currentTime + 0.01);
|
|
destination = null;
|
|
};
|
|
|
|
let workletsLoading;
|
|
function loadWorklets() {
|
|
if (workletsLoading) {
|
|
return workletsLoading;
|
|
}
|
|
workletsLoading = getAudioContext().audioWorklet.addModule(workletsUrl);
|
|
return workletsLoading;
|
|
}
|
|
|
|
function getWorklet(ac, processor, params) {
|
|
const node = new AudioWorkletNode(ac, processor);
|
|
Object.entries(params).forEach(([key, value]) => {
|
|
node.parameters.get(key).value = value;
|
|
});
|
|
return node;
|
|
}
|
|
|
|
// this function should be called on first user interaction (to avoid console warning)
|
|
export async function initAudio() {
|
|
if (typeof window !== 'undefined') {
|
|
try {
|
|
await getAudioContext().resume();
|
|
await loadWorklets();
|
|
} catch (err) {
|
|
console.warn('could not load AudioWorklet effects coarse, crush and shape', err);
|
|
}
|
|
}
|
|
}
|
|
|
|
export async function initAudioOnFirstClick() {
|
|
return new Promise((resolve) => {
|
|
document.addEventListener('click', async function listener() {
|
|
await initAudio();
|
|
resolve();
|
|
document.removeEventListener('click', listener);
|
|
});
|
|
});
|
|
}
|
|
|
|
let delays = {};
|
|
function getDelay(orbit, delaytime, delayfeedback, t) {
|
|
if (!delays[orbit]) {
|
|
const ac = getAudioContext();
|
|
const dly = ac.createFeedbackDelay(1, delaytime, delayfeedback);
|
|
dly.start?.(t); // for some reason, this throws when audion extension is installed..
|
|
dly.connect(getDestination());
|
|
delays[orbit] = dly;
|
|
}
|
|
delays[orbit].delayTime.value !== delaytime && delays[orbit].delayTime.setValueAtTime(delaytime, t);
|
|
delays[orbit].feedback.value !== delayfeedback && delays[orbit].feedback.setValueAtTime(delayfeedback, t);
|
|
return delays[orbit];
|
|
}
|
|
|
|
let reverbs = {};
|
|
function getReverb(orbit, duration = 2) {
|
|
if (!reverbs[orbit]) {
|
|
const ac = getAudioContext();
|
|
const reverb = ac.createReverb(duration);
|
|
reverb.connect(getDestination());
|
|
reverbs[orbit] = reverb;
|
|
}
|
|
if (reverbs[orbit].duration !== duration) {
|
|
reverbs[orbit] = reverbs[orbit].setDuration(duration);
|
|
reverbs[orbit].duration = duration;
|
|
}
|
|
return reverbs[orbit];
|
|
}
|
|
|
|
function effectSend(input, effect, wet) {
|
|
const send = gainNode(wet);
|
|
input.connect(send);
|
|
send.connect(effect);
|
|
return send;
|
|
}
|
|
|
|
// export const webaudioOutput = async (t, hap, ct, cps) => {
|
|
export const webaudioOutput = async (hap, deadline, hapDuration, cps) => {
|
|
const ac = getAudioContext();
|
|
hap.ensureObjectValue();
|
|
|
|
// calculate absolute time
|
|
let t = ac.currentTime + deadline;
|
|
// destructure value
|
|
let {
|
|
s = 'triangle',
|
|
bank,
|
|
source,
|
|
gain = 0.8,
|
|
// low pass
|
|
cutoff,
|
|
resonance = 1,
|
|
// high pass
|
|
hcutoff,
|
|
hresonance = 1,
|
|
// band pass
|
|
bandf,
|
|
bandq = 1,
|
|
//
|
|
coarse,
|
|
crush,
|
|
shape,
|
|
pan,
|
|
vowel,
|
|
delay = 0,
|
|
delayfeedback = 0.5,
|
|
delaytime = 0.25,
|
|
orbit = 1,
|
|
room,
|
|
size = 2,
|
|
} = hap.value;
|
|
const { velocity = 1 } = hap.context;
|
|
gain *= velocity; // legacy fix for velocity
|
|
// the chain will hold all audio nodes that connect to each other
|
|
const chain = [];
|
|
if (bank && s) {
|
|
s = `${bank}_${s}`;
|
|
}
|
|
if (soundMap.has(s)) {
|
|
const node = await soundMap.get(s)({ hap, t, deadline, duration: hapDuration, cps });
|
|
chain.push(node);
|
|
} else if (source) {
|
|
chain.push(source({ hap, t, deadline, duration: hapDuration, cps }));
|
|
} else {
|
|
throw new Error(`sound ${s} not found! Is it loaded?`);
|
|
}
|
|
|
|
// gain stage
|
|
chain.push(gainNode(gain));
|
|
|
|
// filters
|
|
cutoff !== undefined && chain.push(getFilter('lowpass', cutoff, resonance));
|
|
hcutoff !== undefined && chain.push(getFilter('highpass', hcutoff, hresonance));
|
|
bandf !== undefined && chain.push(getFilter('bandpass', bandf, bandq));
|
|
vowel !== undefined && chain.push(ac.createVowelFilter(vowel));
|
|
|
|
// effects
|
|
coarse !== undefined && chain.push(getWorklet(ac, 'coarse-processor', { coarse }));
|
|
crush !== undefined && chain.push(getWorklet(ac, 'crush-processor', { crush }));
|
|
shape !== undefined && chain.push(getWorklet(ac, 'shape-processor', { shape }));
|
|
|
|
// panning
|
|
if (pan !== undefined) {
|
|
const panner = ac.createStereoPanner();
|
|
panner.pan.value = 2 * pan - 1;
|
|
chain.push(panner);
|
|
}
|
|
|
|
// last gain
|
|
const post = gainNode(1);
|
|
chain.push(post);
|
|
post.connect(getDestination());
|
|
|
|
// delay
|
|
let delaySend;
|
|
if (delay > 0 && delaytime > 0 && delayfeedback > 0) {
|
|
const delyNode = getDelay(orbit, delaytime, delayfeedback, t);
|
|
delaySend = effectSend(post, delyNode, delay);
|
|
}
|
|
// reverb
|
|
let reverbSend;
|
|
if (room > 0 && size > 0) {
|
|
const reverbNode = getReverb(orbit, size);
|
|
reverbSend = effectSend(post, reverbNode, room);
|
|
}
|
|
|
|
// connect chain elements together
|
|
chain.slice(1).reduce((last, current) => last.connect(current), chain[0]);
|
|
|
|
// disconnect all nodes when source node has ended:
|
|
chain[0].onended = () => chain.concat([delaySend, reverbSend]).forEach((n) => n?.disconnect());
|
|
};
|
|
|
|
export const webaudioOutputTrigger = (t, hap, ct, cps) => webaudioOutput(hap, t - ct, hap.duration / cps, cps);
|
|
|
|
Pattern.prototype.webaudio = function () {
|
|
// TODO: refactor (t, hap, ct, cps) to (hap, deadline, duration) ?
|
|
return this.onTrigger(webaudioOutputTrigger);
|
|
};
|