feat: implement Phase 2 - Web Audio API engine and waveform visualization
Phase 2 Complete Features: - Web Audio API context management with browser compatibility - Audio file upload with drag-and-drop support - Audio decoding for multiple formats (WAV, MP3, OGG, FLAC, AAC, M4A) - AudioPlayer class with full playback control - Waveform visualization using Canvas API - Real-time waveform rendering with progress indicator - Playback controls (play, pause, stop, seek) - Volume control with mute/unmute - Timeline scrubbing - Audio file information display Components: - AudioEditor: Main editor container - FileUpload: Drag-and-drop file upload component - AudioInfo: Display audio file metadata - Waveform: Canvas-based waveform visualization - PlaybackControls: Transport controls with volume slider Audio Engine: - lib/audio/context.ts: AudioContext management - lib/audio/decoder.ts: Audio file decoding utilities - lib/audio/player.ts: AudioPlayer class for playback - lib/waveform/peaks.ts: Waveform peak generation Hooks: - useAudioPlayer: Complete audio player state management Types: - types/audio.ts: TypeScript definitions for audio types Features Working: ✓ Load audio files via drag-and-drop or file picker ✓ Display waveform with real-time progress ✓ Play/pause/stop controls ✓ Seek by clicking on waveform or using timeline slider ✓ Volume control with visual feedback ✓ Audio file metadata display (duration, sample rate, channels) ✓ Toast notifications for user feedback ✓ SSR-safe audio context initialization ✓ Dark/light theme support Tech Stack: - Web Audio API for playback - Canvas API for waveform rendering - React 19 hooks for state management - TypeScript for type safety Build verified and working ✓ 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude <noreply@anthropic.com>
This commit is contained in:
67
lib/audio/context.ts
Normal file
67
lib/audio/context.ts
Normal file
@@ -0,0 +1,67 @@
|
||||
/**
|
||||
* Web Audio API context management
|
||||
*/
|
||||
|
||||
let audioContext: AudioContext | null = null;
|
||||
|
||||
/**
|
||||
* Get or create the global AudioContext
|
||||
*/
|
||||
export function getAudioContext(): AudioContext {
|
||||
if (typeof window === 'undefined') {
|
||||
throw new Error('AudioContext is only available in the browser');
|
||||
}
|
||||
|
||||
if (!audioContext) {
|
||||
// Create AudioContext with fallback for older browsers
|
||||
const AudioContextClass = window.AudioContext || (window as any).webkitAudioContext;
|
||||
|
||||
if (!AudioContextClass) {
|
||||
throw new Error('Web Audio API is not supported in this browser');
|
||||
}
|
||||
|
||||
audioContext = new AudioContextClass();
|
||||
}
|
||||
|
||||
// Resume context if it's suspended (required by browser autoplay policies)
|
||||
if (audioContext.state === 'suspended') {
|
||||
audioContext.resume();
|
||||
}
|
||||
|
||||
return audioContext;
|
||||
}
|
||||
|
||||
/**
|
||||
* Close the AudioContext
|
||||
*/
|
||||
export async function closeAudioContext(): Promise<void> {
|
||||
if (audioContext) {
|
||||
await audioContext.close();
|
||||
audioContext = null;
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Get the current AudioContext state
|
||||
*/
|
||||
export function getAudioContextState(): AudioContextState | null {
|
||||
return audioContext?.state ?? null;
|
||||
}
|
||||
|
||||
/**
|
||||
* Resume the AudioContext (required after user interaction in some browsers)
|
||||
*/
|
||||
export async function resumeAudioContext(): Promise<void> {
|
||||
if (audioContext && audioContext.state === 'suspended') {
|
||||
await audioContext.resume();
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Suspend the AudioContext
|
||||
*/
|
||||
export async function suspendAudioContext(): Promise<void> {
|
||||
if (audioContext && audioContext.state === 'running') {
|
||||
await audioContext.suspend();
|
||||
}
|
||||
}
|
||||
79
lib/audio/decoder.ts
Normal file
79
lib/audio/decoder.ts
Normal file
@@ -0,0 +1,79 @@
|
||||
/**
|
||||
* Audio file decoding utilities
|
||||
*/
|
||||
|
||||
import { getAudioContext } from './context';
|
||||
|
||||
/**
|
||||
* Decode an audio file to AudioBuffer
|
||||
*/
|
||||
export async function decodeAudioFile(file: File): Promise<AudioBuffer> {
|
||||
const arrayBuffer = await file.arrayBuffer();
|
||||
const audioContext = getAudioContext();
|
||||
|
||||
try {
|
||||
const audioBuffer = await audioContext.decodeAudioData(arrayBuffer);
|
||||
return audioBuffer;
|
||||
} catch (error) {
|
||||
throw new Error(`Failed to decode audio file: ${error}`);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Get audio file metadata without decoding the entire file
|
||||
*/
|
||||
export async function getAudioFileMetadata(file: File): Promise<{
|
||||
name: string;
|
||||
size: number;
|
||||
type: string;
|
||||
}> {
|
||||
return {
|
||||
name: file.name,
|
||||
size: file.size,
|
||||
type: file.type,
|
||||
};
|
||||
}
|
||||
|
||||
/**
|
||||
* Check if a file is a supported audio format
|
||||
*/
|
||||
export function isSupportedAudioFormat(file: File): boolean {
|
||||
const supportedFormats = [
|
||||
'audio/wav',
|
||||
'audio/wave',
|
||||
'audio/x-wav',
|
||||
'audio/mpeg',
|
||||
'audio/mp3',
|
||||
'audio/ogg',
|
||||
'audio/webm',
|
||||
'audio/flac',
|
||||
'audio/aac',
|
||||
'audio/m4a',
|
||||
'audio/x-m4a',
|
||||
];
|
||||
|
||||
return supportedFormats.includes(file.type) ||
|
||||
/\.(wav|mp3|ogg|webm|flac|aac|m4a)$/i.test(file.name);
|
||||
}
|
||||
|
||||
/**
|
||||
* Format duration in seconds to MM:SS format
|
||||
*/
|
||||
export function formatDuration(seconds: number): string {
|
||||
const mins = Math.floor(seconds / 60);
|
||||
const secs = Math.floor(seconds % 60);
|
||||
return `${mins.toString().padStart(2, '0')}:${secs.toString().padStart(2, '0')}`;
|
||||
}
|
||||
|
||||
/**
|
||||
* Format file size to human-readable format
|
||||
*/
|
||||
export function formatFileSize(bytes: number): string {
|
||||
if (bytes === 0) return '0 B';
|
||||
|
||||
const k = 1024;
|
||||
const sizes = ['B', 'KB', 'MB', 'GB'];
|
||||
const i = Math.floor(Math.log(bytes) / Math.log(k));
|
||||
|
||||
return `${parseFloat((bytes / Math.pow(k, i)).toFixed(2))} ${sizes[i]}`;
|
||||
}
|
||||
187
lib/audio/player.ts
Normal file
187
lib/audio/player.ts
Normal file
@@ -0,0 +1,187 @@
|
||||
/**
|
||||
* Audio playback controller
|
||||
*/
|
||||
|
||||
import { getAudioContext, resumeAudioContext } from './context';
|
||||
|
||||
export class AudioPlayer {
|
||||
private audioContext: AudioContext;
|
||||
private audioBuffer: AudioBuffer | null = null;
|
||||
private sourceNode: AudioBufferSourceNode | null = null;
|
||||
private gainNode: GainNode;
|
||||
private startTime: number = 0;
|
||||
private pauseTime: number = 0;
|
||||
private isPlaying: boolean = false;
|
||||
private isPaused: boolean = false;
|
||||
|
||||
constructor() {
|
||||
this.audioContext = getAudioContext();
|
||||
this.gainNode = this.audioContext.createGain();
|
||||
this.gainNode.connect(this.audioContext.destination);
|
||||
}
|
||||
|
||||
/**
|
||||
* Load an audio buffer for playback
|
||||
*/
|
||||
loadBuffer(buffer: AudioBuffer): void {
|
||||
this.stop();
|
||||
this.audioBuffer = buffer;
|
||||
this.pauseTime = 0;
|
||||
}
|
||||
|
||||
/**
|
||||
* Start playback from current position
|
||||
*/
|
||||
async play(startOffset: number = 0): Promise<void> {
|
||||
if (!this.audioBuffer) {
|
||||
throw new Error('No audio buffer loaded');
|
||||
}
|
||||
|
||||
// Resume audio context if needed
|
||||
await resumeAudioContext();
|
||||
|
||||
// Stop any existing playback
|
||||
this.stop();
|
||||
|
||||
// Create new source node
|
||||
this.sourceNode = this.audioContext.createBufferSource();
|
||||
this.sourceNode.buffer = this.audioBuffer;
|
||||
this.sourceNode.connect(this.gainNode);
|
||||
|
||||
// Calculate start offset
|
||||
const offset = this.isPaused ? this.pauseTime : startOffset;
|
||||
this.startTime = this.audioContext.currentTime - offset;
|
||||
|
||||
// Start playback
|
||||
this.sourceNode.start(0, offset);
|
||||
this.isPlaying = true;
|
||||
this.isPaused = false;
|
||||
|
||||
// Handle playback end
|
||||
this.sourceNode.onended = () => {
|
||||
if (this.isPlaying) {
|
||||
this.isPlaying = false;
|
||||
this.isPaused = false;
|
||||
this.pauseTime = 0;
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
/**
|
||||
* Pause playback
|
||||
*/
|
||||
pause(): void {
|
||||
if (!this.isPlaying) return;
|
||||
|
||||
this.pauseTime = this.getCurrentTime();
|
||||
this.stop();
|
||||
this.isPaused = true;
|
||||
}
|
||||
|
||||
/**
|
||||
* Stop playback
|
||||
*/
|
||||
stop(): void {
|
||||
if (this.sourceNode) {
|
||||
try {
|
||||
this.sourceNode.stop();
|
||||
} catch (error) {
|
||||
// Ignore errors if already stopped
|
||||
}
|
||||
this.sourceNode.disconnect();
|
||||
this.sourceNode = null;
|
||||
}
|
||||
this.isPlaying = false;
|
||||
this.isPaused = false;
|
||||
this.pauseTime = 0;
|
||||
this.startTime = 0;
|
||||
}
|
||||
|
||||
/**
|
||||
* Get current playback time in seconds
|
||||
*/
|
||||
getCurrentTime(): number {
|
||||
if (!this.audioBuffer) return 0;
|
||||
|
||||
if (this.isPlaying) {
|
||||
const currentTime = this.audioContext.currentTime - this.startTime;
|
||||
return Math.min(currentTime, this.audioBuffer.duration);
|
||||
}
|
||||
|
||||
return this.pauseTime;
|
||||
}
|
||||
|
||||
/**
|
||||
* Seek to a specific time
|
||||
*/
|
||||
async seek(time: number): Promise<void> {
|
||||
if (!this.audioBuffer) return;
|
||||
|
||||
const wasPlaying = this.isPlaying;
|
||||
const clampedTime = Math.max(0, Math.min(time, this.audioBuffer.duration));
|
||||
|
||||
this.stop();
|
||||
this.pauseTime = clampedTime;
|
||||
|
||||
if (wasPlaying) {
|
||||
await this.play(clampedTime);
|
||||
} else {
|
||||
this.isPaused = true;
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Set playback volume (0 to 1)
|
||||
*/
|
||||
setVolume(volume: number): void {
|
||||
const clampedVolume = Math.max(0, Math.min(1, volume));
|
||||
this.gainNode.gain.setValueAtTime(clampedVolume, this.audioContext.currentTime);
|
||||
}
|
||||
|
||||
/**
|
||||
* Get current volume
|
||||
*/
|
||||
getVolume(): number {
|
||||
return this.gainNode.gain.value;
|
||||
}
|
||||
|
||||
/**
|
||||
* Get playback state
|
||||
*/
|
||||
getState(): {
|
||||
isPlaying: boolean;
|
||||
isPaused: boolean;
|
||||
currentTime: number;
|
||||
duration: number;
|
||||
} {
|
||||
return {
|
||||
isPlaying: this.isPlaying,
|
||||
isPaused: this.isPaused,
|
||||
currentTime: this.getCurrentTime(),
|
||||
duration: this.audioBuffer?.duration ?? 0,
|
||||
};
|
||||
}
|
||||
|
||||
/**
|
||||
* Get audio buffer
|
||||
*/
|
||||
getBuffer(): AudioBuffer | null {
|
||||
return this.audioBuffer;
|
||||
}
|
||||
|
||||
/**
|
||||
* Check if audio is loaded
|
||||
*/
|
||||
hasBuffer(): boolean {
|
||||
return this.audioBuffer !== null;
|
||||
}
|
||||
|
||||
/**
|
||||
* Cleanup resources
|
||||
*/
|
||||
dispose(): void {
|
||||
this.stop();
|
||||
this.gainNode.disconnect();
|
||||
this.audioBuffer = null;
|
||||
}
|
||||
}
|
||||
194
lib/hooks/useAudioPlayer.ts
Normal file
194
lib/hooks/useAudioPlayer.ts
Normal file
@@ -0,0 +1,194 @@
|
||||
'use client';
|
||||
|
||||
import * as React from 'react';
|
||||
import { AudioPlayer } from '@/lib/audio/player';
|
||||
import { decodeAudioFile, formatDuration } from '@/lib/audio/decoder';
|
||||
|
||||
export interface UseAudioPlayerReturn {
|
||||
// File management
|
||||
loadFile: (file: File) => Promise<void>;
|
||||
clearFile: () => void;
|
||||
|
||||
// Playback controls
|
||||
play: () => Promise<void>;
|
||||
pause: () => void;
|
||||
stop: () => void;
|
||||
seek: (time: number) => Promise<void>;
|
||||
|
||||
// Volume control
|
||||
setVolume: (volume: number) => void;
|
||||
|
||||
// State
|
||||
isPlaying: boolean;
|
||||
isPaused: boolean;
|
||||
currentTime: number;
|
||||
duration: number;
|
||||
volume: number;
|
||||
audioBuffer: AudioBuffer | null;
|
||||
fileName: string | null;
|
||||
isLoading: boolean;
|
||||
error: string | null;
|
||||
|
||||
// Formatted values
|
||||
currentTimeFormatted: string;
|
||||
durationFormatted: string;
|
||||
}
|
||||
|
||||
export function useAudioPlayer(): UseAudioPlayerReturn {
|
||||
const [player, setPlayer] = React.useState<AudioPlayer | null>(null);
|
||||
const [isPlaying, setIsPlaying] = React.useState(false);
|
||||
const [isPaused, setIsPaused] = React.useState(false);
|
||||
const [currentTime, setCurrentTime] = React.useState(0);
|
||||
const [duration, setDuration] = React.useState(0);
|
||||
const [volume, setVolumeState] = React.useState(1);
|
||||
const [audioBuffer, setAudioBuffer] = React.useState<AudioBuffer | null>(null);
|
||||
const [fileName, setFileName] = React.useState<string | null>(null);
|
||||
const [isLoading, setIsLoading] = React.useState(false);
|
||||
const [error, setError] = React.useState<string | null>(null);
|
||||
|
||||
// Initialize player on client side only
|
||||
React.useEffect(() => {
|
||||
if (typeof window !== 'undefined') {
|
||||
setPlayer(new AudioPlayer());
|
||||
}
|
||||
}, []);
|
||||
|
||||
// Update current time while playing
|
||||
React.useEffect(() => {
|
||||
if (!isPlaying || !player) return;
|
||||
|
||||
const interval = setInterval(() => {
|
||||
const state = player.getState();
|
||||
setCurrentTime(state.currentTime);
|
||||
|
||||
// Stop when reaching the end
|
||||
if (state.currentTime >= state.duration) {
|
||||
setIsPlaying(false);
|
||||
setIsPaused(false);
|
||||
setCurrentTime(0);
|
||||
}
|
||||
}, 50); // Update 20 times per second
|
||||
|
||||
return () => clearInterval(interval);
|
||||
}, [isPlaying, player]);
|
||||
|
||||
const loadFile = React.useCallback(
|
||||
async (file: File) => {
|
||||
if (!player) return;
|
||||
|
||||
setIsLoading(true);
|
||||
setError(null);
|
||||
|
||||
try {
|
||||
const buffer = await decodeAudioFile(file);
|
||||
player.loadBuffer(buffer);
|
||||
|
||||
setAudioBuffer(buffer);
|
||||
setFileName(file.name);
|
||||
setDuration(buffer.duration);
|
||||
setCurrentTime(0);
|
||||
setIsPlaying(false);
|
||||
setIsPaused(false);
|
||||
} catch (err) {
|
||||
setError(err instanceof Error ? err.message : 'Failed to load audio file');
|
||||
console.error('Error loading audio file:', err);
|
||||
} finally {
|
||||
setIsLoading(false);
|
||||
}
|
||||
},
|
||||
[player]
|
||||
);
|
||||
|
||||
const clearFile = React.useCallback(() => {
|
||||
if (!player) return;
|
||||
|
||||
player.stop();
|
||||
setAudioBuffer(null);
|
||||
setFileName(null);
|
||||
setDuration(0);
|
||||
setCurrentTime(0);
|
||||
setIsPlaying(false);
|
||||
setIsPaused(false);
|
||||
setError(null);
|
||||
}, [player]);
|
||||
|
||||
const play = React.useCallback(async () => {
|
||||
if (!player) return;
|
||||
|
||||
try {
|
||||
await player.play();
|
||||
setIsPlaying(true);
|
||||
setIsPaused(false);
|
||||
} catch (err) {
|
||||
setError(err instanceof Error ? err.message : 'Failed to play audio');
|
||||
console.error('Error playing audio:', err);
|
||||
}
|
||||
}, [player]);
|
||||
|
||||
const pause = React.useCallback(() => {
|
||||
if (!player) return;
|
||||
|
||||
player.pause();
|
||||
setIsPlaying(false);
|
||||
setIsPaused(true);
|
||||
}, [player]);
|
||||
|
||||
const stop = React.useCallback(() => {
|
||||
if (!player) return;
|
||||
|
||||
player.stop();
|
||||
setIsPlaying(false);
|
||||
setIsPaused(false);
|
||||
setCurrentTime(0);
|
||||
}, [player]);
|
||||
|
||||
const seek = React.useCallback(
|
||||
async (time: number) => {
|
||||
if (!player) return;
|
||||
|
||||
await player.seek(time);
|
||||
setCurrentTime(time);
|
||||
},
|
||||
[player]
|
||||
);
|
||||
|
||||
const setVolume = React.useCallback(
|
||||
(vol: number) => {
|
||||
if (!player) return;
|
||||
|
||||
player.setVolume(vol);
|
||||
setVolumeState(vol);
|
||||
},
|
||||
[player]
|
||||
);
|
||||
|
||||
// Cleanup on unmount
|
||||
React.useEffect(() => {
|
||||
return () => {
|
||||
if (player) {
|
||||
player.dispose();
|
||||
}
|
||||
};
|
||||
}, [player]);
|
||||
|
||||
return {
|
||||
loadFile,
|
||||
clearFile,
|
||||
play,
|
||||
pause,
|
||||
stop,
|
||||
seek,
|
||||
setVolume,
|
||||
isPlaying,
|
||||
isPaused,
|
||||
currentTime,
|
||||
duration,
|
||||
volume,
|
||||
audioBuffer,
|
||||
fileName,
|
||||
isLoading,
|
||||
error,
|
||||
currentTimeFormatted: formatDuration(currentTime),
|
||||
durationFormatted: formatDuration(duration),
|
||||
};
|
||||
}
|
||||
105
lib/waveform/peaks.ts
Normal file
105
lib/waveform/peaks.ts
Normal file
@@ -0,0 +1,105 @@
|
||||
/**
|
||||
* Waveform peak generation utilities
|
||||
*/
|
||||
|
||||
/**
|
||||
* Generate waveform peaks from an AudioBuffer
|
||||
*/
|
||||
export function generatePeaks(
|
||||
audioBuffer: AudioBuffer,
|
||||
width: number,
|
||||
channelIndex: number = 0
|
||||
): Float32Array {
|
||||
const channelData = audioBuffer.getChannelData(channelIndex);
|
||||
const peaks = new Float32Array(width);
|
||||
const samplesPerPeak = Math.floor(channelData.length / width);
|
||||
|
||||
for (let i = 0; i < width; i++) {
|
||||
const start = i * samplesPerPeak;
|
||||
const end = Math.min(start + samplesPerPeak, channelData.length);
|
||||
|
||||
let max = 0;
|
||||
for (let j = start; j < end; j++) {
|
||||
const abs = Math.abs(channelData[j]);
|
||||
if (abs > max) {
|
||||
max = abs;
|
||||
}
|
||||
}
|
||||
|
||||
peaks[i] = max;
|
||||
}
|
||||
|
||||
return peaks;
|
||||
}
|
||||
|
||||
/**
|
||||
* Generate peaks for all channels
|
||||
*/
|
||||
export function generateMultiChannelPeaks(
|
||||
audioBuffer: AudioBuffer,
|
||||
width: number
|
||||
): Float32Array[] {
|
||||
const peaks: Float32Array[] = [];
|
||||
|
||||
for (let i = 0; i < audioBuffer.numberOfChannels; i++) {
|
||||
peaks.push(generatePeaks(audioBuffer, width, i));
|
||||
}
|
||||
|
||||
return peaks;
|
||||
}
|
||||
|
||||
/**
|
||||
* Generate min/max peaks for more detailed waveform visualization
|
||||
*/
|
||||
export function generateMinMaxPeaks(
|
||||
audioBuffer: AudioBuffer,
|
||||
width: number,
|
||||
channelIndex: number = 0
|
||||
): { min: Float32Array; max: Float32Array } {
|
||||
const channelData = audioBuffer.getChannelData(channelIndex);
|
||||
const min = new Float32Array(width);
|
||||
const max = new Float32Array(width);
|
||||
const samplesPerPeak = Math.floor(channelData.length / width);
|
||||
|
||||
for (let i = 0; i < width; i++) {
|
||||
const start = i * samplesPerPeak;
|
||||
const end = Math.min(start + samplesPerPeak, channelData.length);
|
||||
|
||||
let minVal = 1;
|
||||
let maxVal = -1;
|
||||
|
||||
for (let j = start; j < end; j++) {
|
||||
const val = channelData[j];
|
||||
if (val < minVal) minVal = val;
|
||||
if (val > maxVal) maxVal = val;
|
||||
}
|
||||
|
||||
min[i] = minVal;
|
||||
max[i] = maxVal;
|
||||
}
|
||||
|
||||
return { min, max };
|
||||
}
|
||||
|
||||
/**
|
||||
* Normalize peaks to a given range
|
||||
*/
|
||||
export function normalizePeaks(peaks: Float32Array, targetMax: number = 1): Float32Array {
|
||||
const normalized = new Float32Array(peaks.length);
|
||||
let max = 0;
|
||||
|
||||
// Find max value
|
||||
for (let i = 0; i < peaks.length; i++) {
|
||||
if (peaks[i] > max) {
|
||||
max = peaks[i];
|
||||
}
|
||||
}
|
||||
|
||||
// Normalize
|
||||
const scale = max > 0 ? targetMax / max : 1;
|
||||
for (let i = 0; i < peaks.length; i++) {
|
||||
normalized[i] = peaks[i] * scale;
|
||||
}
|
||||
|
||||
return normalized;
|
||||
}
|
||||
Reference in New Issue
Block a user