Set up effect callbacks to swap audio buffers at end of clips
This commit is contained in:
parent
9867c21111
commit
3da297315f
130
pages/index.tsx
130
pages/index.tsx
|
@ -1,5 +1,5 @@
|
||||||
import Head from "next/head";
|
import Head from "next/head";
|
||||||
import { useEffect, useState } from "react";
|
import { useEffect, useState, useRef } from "react";
|
||||||
|
|
||||||
import ThreeCanvas from "../components/ThreeCanvas";
|
import ThreeCanvas from "../components/ThreeCanvas";
|
||||||
import PromptPanel from "../components/PromptPanel";
|
import PromptPanel from "../components/PromptPanel";
|
||||||
|
@ -11,12 +11,15 @@ import { InferenceResult, PromptInput } from "../types";
|
||||||
import * as Tone from "tone";
|
import * as Tone from "tone";
|
||||||
|
|
||||||
const defaultPromptInputs = [
|
const defaultPromptInputs = [
|
||||||
{ prompt: "A jazz pianist playing a classical concerto"},
|
{ prompt: "A jazz pianist playing a classical concerto" },
|
||||||
{ prompt: "Taylor Swift singing with a tropical beat"},
|
{ prompt: "Taylor Swift singing with a tropical beat" },
|
||||||
{ prompt: "A typewriter in the bahamas"},
|
{ prompt: "A typewriter in the bahamas" },
|
||||||
{ prompt: "Justin Bieber anger rap"},
|
{ prompt: "Justin Bieber anger rap" },
|
||||||
{ prompt: "New york city rap, with a dust storm, cinematic score, dramatic, composition"},
|
{
|
||||||
{ prompt: "Jack Johnson playing a harmonica in the 1920s"},
|
prompt:
|
||||||
|
"New york city rap, with a dust storm, cinematic score, dramatic, composition",
|
||||||
|
},
|
||||||
|
{ prompt: "Jack Johnson playing a harmonica in the 1920s" },
|
||||||
];
|
];
|
||||||
|
|
||||||
const defaultInferenceResults = [
|
const defaultInferenceResults = [
|
||||||
|
@ -30,6 +33,16 @@ const defaultInferenceResults = [
|
||||||
audio: "rap_sample.mp3",
|
audio: "rap_sample.mp3",
|
||||||
counter: 0,
|
counter: 0,
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
input: {
|
||||||
|
alpha: 0.0,
|
||||||
|
start: defaultPromptInputs[0],
|
||||||
|
end: defaultPromptInputs[1],
|
||||||
|
},
|
||||||
|
image: "pop_sample.jpg",
|
||||||
|
audio: "pop_sample.mp3",
|
||||||
|
counter: 1,
|
||||||
|
},
|
||||||
];
|
];
|
||||||
|
|
||||||
// TODO(hayk): Do this as soon as sample comes back
|
// TODO(hayk): Do this as soon as sample comes back
|
||||||
|
@ -50,34 +63,39 @@ export default function Home() {
|
||||||
|
|
||||||
const [tonePlayer, setTonePlayer] = useState<Tone.Player>(null);
|
const [tonePlayer, setTonePlayer] = useState<Tone.Player>(null);
|
||||||
|
|
||||||
|
const [numClipsPlayed, setNumClipsPlayed] = useState(0);
|
||||||
|
const [prevNumClipsPlayed, setPrevNumClipsPlayed] = useState(0);
|
||||||
|
|
||||||
|
const [resultCounter, setResultCounter] = useState(0);
|
||||||
|
|
||||||
|
// On load, create a player synced to the tone transport
|
||||||
useEffect(() => {
|
useEffect(() => {
|
||||||
// HACK(hayk): Kill
|
const audioUrl = inferenceResults[0].audio;
|
||||||
if (tonePlayer) {
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (inferenceResults.length == 0) {
|
const player = new Tone.Player(audioUrl, () => {
|
||||||
return;
|
console.log("Created player.");
|
||||||
}
|
|
||||||
console.log(inferenceResults);
|
|
||||||
|
|
||||||
const player = new Tone.Player(
|
|
||||||
inferenceResults[inferenceResults.length - 1].audio,
|
|
||||||
() => {
|
|
||||||
console.log("New player loaded.");
|
|
||||||
|
|
||||||
|
player.loop = true;
|
||||||
player.sync().start(0);
|
player.sync().start(0);
|
||||||
|
|
||||||
// if (tonePlayer) {
|
// Set up a callback to increment numClipsPlayed at the edge of each clip
|
||||||
// tonePlayer.stop();
|
const bufferLength = player.sampleTime * player.buffer.length;
|
||||||
// tonePlayer.dispose();
|
Tone.Transport.scheduleRepeat((time) => {
|
||||||
// }
|
console.log(
|
||||||
setTonePlayer(player);
|
"Edge of clip, t = ",
|
||||||
}
|
Tone.Transport.getSecondsAtTime(time)
|
||||||
).toDestination();
|
);
|
||||||
player.loop = true;
|
setNumClipsPlayed((n) => n + 1);
|
||||||
}, [inferenceResults]);
|
}, bufferLength);
|
||||||
|
|
||||||
|
setTonePlayer(player);
|
||||||
|
|
||||||
|
// Make further load callbacks do nothing.
|
||||||
|
player.buffer.onload = () => {};
|
||||||
|
}).toDestination();
|
||||||
|
}, []);
|
||||||
|
|
||||||
|
// On play/pause button, play/pause the audio with the tone transport
|
||||||
useEffect(() => {
|
useEffect(() => {
|
||||||
if (!paused) {
|
if (!paused) {
|
||||||
console.log("Play");
|
console.log("Play");
|
||||||
|
@ -98,14 +116,58 @@ export default function Home() {
|
||||||
}
|
}
|
||||||
}, [paused, tonePlayer]);
|
}, [paused, tonePlayer]);
|
||||||
|
|
||||||
|
useEffect(() => {
|
||||||
|
if (numClipsPlayed == prevNumClipsPlayed) {
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
const maxResultCounter = Math.max(
|
||||||
|
...inferenceResults.map((r) => r.counter)
|
||||||
|
);
|
||||||
|
|
||||||
|
if (maxResultCounter <= resultCounter) {
|
||||||
|
console.info("not picking a new result, none available");
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
const result = inferenceResults.find(
|
||||||
|
(r: InferenceResult) => r.counter == resultCounter
|
||||||
|
);
|
||||||
|
|
||||||
|
console.log("Incrementing result counter ", resultCounter, result);
|
||||||
|
setResultCounter((c) => c + 1);
|
||||||
|
|
||||||
|
tonePlayer.load(result.audio).then(() => {
|
||||||
|
console.log("Loaded new: ", result.audio);
|
||||||
|
|
||||||
|
// Re-jigger the transport so it stops playing old buffers. It seems like this doesn't
|
||||||
|
// introduce a gap, but watch out for that.
|
||||||
|
Tone.Transport.pause();
|
||||||
|
if (!paused) {
|
||||||
|
Tone.Transport.start();
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
setPrevNumClipsPlayed(numClipsPlayed);
|
||||||
|
}, [
|
||||||
|
numClipsPlayed,
|
||||||
|
prevNumClipsPlayed,
|
||||||
|
resultCounter,
|
||||||
|
inferenceResults,
|
||||||
|
paused,
|
||||||
|
tonePlayer,
|
||||||
|
]);
|
||||||
|
|
||||||
// /////////////
|
// /////////////
|
||||||
|
|
||||||
|
// Request a new inference prompt at a regular interval
|
||||||
|
// TODO(hayk): Rewrite this to request more frequently and max out
|
||||||
useEffect(() => {
|
useEffect(() => {
|
||||||
console.log("setInterval");
|
console.log("setInterval");
|
||||||
setInterval(() => {
|
setInterval(() => {
|
||||||
setInferenceResults((prevResults) => {
|
setInferenceResults((prevResults) => {
|
||||||
const lastResult = prevResults[prevResults.length - 1];
|
const lastResult = prevResults[prevResults.length - 2];
|
||||||
const newResult = { ...lastResult, counter: lastResult.counter + 1 };
|
const newResult = { ...lastResult, counter: lastResult.counter + 2 };
|
||||||
|
|
||||||
let results = [...prevResults, newResult];
|
let results = [...prevResults, newResult];
|
||||||
|
|
||||||
|
@ -113,6 +175,8 @@ export default function Home() {
|
||||||
results = results.slice(1);
|
results = results.slice(1);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
console.log(results);
|
||||||
|
|
||||||
return results;
|
return results;
|
||||||
});
|
});
|
||||||
}, timeout);
|
}, timeout);
|
||||||
|
@ -134,7 +198,9 @@ export default function Home() {
|
||||||
<ThreeCanvas
|
<ThreeCanvas
|
||||||
paused={paused}
|
paused={paused}
|
||||||
getTime={() => Tone.Transport.seconds}
|
getTime={() => Tone.Transport.seconds}
|
||||||
audioLength={tonePlayer ? tonePlayer.sampleTime * tonePlayer.buffer.length : 5}
|
audioLength={
|
||||||
|
tonePlayer ? tonePlayer.sampleTime * tonePlayer.buffer.length : 5
|
||||||
|
}
|
||||||
inferenceResults={inferenceResults}
|
inferenceResults={inferenceResults}
|
||||||
/>
|
/>
|
||||||
</div>
|
</div>
|
||||||
|
|
Binary file not shown.
After Width: | Height: | Size: 204 KiB |
Binary file not shown.
Loading…
Reference in New Issue