mirror of
https://github.com/ChatGPTNextWeb/ChatGPT-Next-Web.git
synced 2025-10-01 15:46:39 +08:00
feat: add arrayBufferToWav utility for audio processing
This commit is contained in:
parent
a90090c162
commit
853212cf13
@ -117,7 +117,7 @@ import { MultimodalContent } from "../client/api";
|
|||||||
|
|
||||||
const localStorage = safeLocalStorage();
|
const localStorage = safeLocalStorage();
|
||||||
import { ClientApi } from "../client/api";
|
import { ClientApi } from "../client/api";
|
||||||
import { createTTSPlayer } from "../utils/audio";
|
import { createTTSPlayer, arrayBufferToWav } from "../utils/audio";
|
||||||
import { MsEdgeTTS, OUTPUT_FORMAT } from "../utils/ms_edge_tts";
|
import { MsEdgeTTS, OUTPUT_FORMAT } from "../utils/ms_edge_tts";
|
||||||
|
|
||||||
const ttsPlayer = createTTSPlayer();
|
const ttsPlayer = createTTSPlayer();
|
||||||
@ -1228,11 +1228,13 @@ function _Chat() {
|
|||||||
}
|
}
|
||||||
setSpeechStatus(true);
|
setSpeechStatus(true);
|
||||||
try {
|
try {
|
||||||
|
const waveFile = arrayBufferToWav(audioBuffer);
|
||||||
|
const audioFile = new Blob([waveFile], { type: "audio/wav" });
|
||||||
|
const url = uploadAudio(audioFile);
|
||||||
|
|
||||||
await ttsPlayer.play(audioBuffer, () => {
|
await ttsPlayer.play(audioBuffer, () => {
|
||||||
setSpeechStatus(false);
|
setSpeechStatus(false);
|
||||||
});
|
});
|
||||||
const audioFile = new Blob([audioBuffer], { type: "audio/wav" });
|
|
||||||
const url = uploadAudio(audioFile);
|
|
||||||
} catch (e) {
|
} catch (e) {
|
||||||
console.error("[OpenAI Speech]", e);
|
console.error("[OpenAI Speech]", e);
|
||||||
showToast(prettyObject(e));
|
showToast(prettyObject(e));
|
||||||
|
@ -43,3 +43,54 @@ export function createTTSPlayer(): TTSPlayer {
|
|||||||
|
|
||||||
return { init, play, stop };
|
return { init, play, stop };
|
||||||
}
|
}
|
||||||
|
|
||||||
|
export function arrayBufferToWav(buffer: ArrayBuffer): ArrayBuffer {
|
||||||
|
const numOfChannels = 1; // Mono
|
||||||
|
const sampleRate = 24000; // 24kHz
|
||||||
|
const bitsPerSample = 16;
|
||||||
|
|
||||||
|
const bytesPerSample = bitsPerSample / 8;
|
||||||
|
const blockAlign = numOfChannels * bytesPerSample;
|
||||||
|
const byteRate = sampleRate * blockAlign;
|
||||||
|
|
||||||
|
// WAV header size is 44 bytes
|
||||||
|
const wavHeaderSize = 44;
|
||||||
|
const dataSize = buffer.byteLength;
|
||||||
|
const totalSize = wavHeaderSize + dataSize;
|
||||||
|
|
||||||
|
const wavBuffer = new ArrayBuffer(totalSize);
|
||||||
|
const view = new DataView(wavBuffer);
|
||||||
|
|
||||||
|
// RIFF chunk descriptor
|
||||||
|
writeString(view, 0, "RIFF");
|
||||||
|
view.setUint32(4, totalSize - 8, true); // File size minus RIFF header
|
||||||
|
writeString(view, 8, "WAVE");
|
||||||
|
|
||||||
|
// FMT sub-chunk
|
||||||
|
writeString(view, 12, "fmt ");
|
||||||
|
view.setUint32(16, 16, true); // Sub-chunk size (16 for PCM)
|
||||||
|
view.setUint16(20, 1, true); // Audio format (1 for PCM)
|
||||||
|
view.setUint16(22, numOfChannels, true); // Number of channels
|
||||||
|
view.setUint32(24, sampleRate, true); // Sample rate
|
||||||
|
view.setUint32(28, byteRate, true); // Byte rate
|
||||||
|
view.setUint16(32, blockAlign, true); // Block align
|
||||||
|
view.setUint16(34, bitsPerSample, true); // Bits per sample
|
||||||
|
|
||||||
|
// Data sub-chunk
|
||||||
|
writeString(view, 36, "data");
|
||||||
|
view.setUint32(40, dataSize, true); // Data size
|
||||||
|
|
||||||
|
// Write the PCM samples
|
||||||
|
const audioData = new Uint8Array(buffer);
|
||||||
|
const wavData = new Uint8Array(wavBuffer);
|
||||||
|
wavData.set(audioData, wavHeaderSize);
|
||||||
|
|
||||||
|
return wavBuffer;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Helper function to write a string to the DataView
|
||||||
|
function writeString(view: DataView, offset: number, string: string) {
|
||||||
|
for (let i = 0; i < string.length; i++) {
|
||||||
|
view.setUint8(offset + i, string.charCodeAt(i));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user