an attempt to implement audioworklets was made. works mostly, but skips
samples
This commit is contained in:
parent
6bc928b5b6
commit
778591d460
59
htdocs/lib/AudioProcessor.js
Normal file
59
htdocs/lib/AudioProcessor.js
Normal file
@ -0,0 +1,59 @@
|
||||
class OwrxAudioProcessor extends AudioWorkletProcessor {
|
||||
constructor(options){
|
||||
super(options);
|
||||
this.maxLength = options.processorOptions.maxLength;
|
||||
this.reduceToLength = options.processorOptions.reduceToLength;
|
||||
this.audio_buffers = [];
|
||||
this.port.addEventListener('message', (m) => {
|
||||
if (typeof(m.data) === 'string') {
|
||||
const json = JSON.parse(m.data);
|
||||
if (json.cmd && json.cmd == 'getBuffers') {
|
||||
this.reportBuffers();
|
||||
}
|
||||
} else {
|
||||
this.audio_buffers.push(new Float32Array(m.data));
|
||||
}
|
||||
});
|
||||
this.port.addEventListener('messageerror', console.error);
|
||||
this.port.start();
|
||||
}
|
||||
process(inputs, outputs, parameters) {
|
||||
//console.time('audio::process');
|
||||
outputs[0].forEach((output) => {
|
||||
let total = 0;
|
||||
while (this.audio_buffers.length) {
|
||||
const b = this.audio_buffers.shift();
|
||||
const newLength = total + b.length;
|
||||
const ol = output.length;
|
||||
// not enough space to fit all data, so splice and put back in the queue
|
||||
if (newLength > ol) {
|
||||
const tokeep = b.slice(0, ol - total);
|
||||
output.set(tokeep, total);
|
||||
const tobuffer = b.slice(ol - total, b.length);
|
||||
this.audio_buffers.unshift(tobuffer);
|
||||
break;
|
||||
} else {
|
||||
output.set(b, total);
|
||||
}
|
||||
total = newLength;
|
||||
}
|
||||
});
|
||||
//console.timeEnd('audio::process');
|
||||
return true;
|
||||
}
|
||||
bufferLength() {
|
||||
return this.audio_buffers.map(function(b){ return b.length; }).reduce(function(a, b){ return a + b; }, 0);
|
||||
}
|
||||
reportBuffers() {
|
||||
var we_have_more_than = (sec) => {
|
||||
return sec * sampleRate < this.bufferLength();
|
||||
};
|
||||
if (we_have_more_than(this.maxLength)) while (we_have_more_than(this.reduceToLength)) {
|
||||
this.audio_buffers.shift();
|
||||
}
|
||||
|
||||
this.port.postMessage(JSON.stringify({buffersize: this.bufferLength()}));
|
||||
}
|
||||
}
|
||||
|
||||
registerProcessor('openwebrx-audio-processor', OwrxAudioProcessor);
|
@ -1544,12 +1544,19 @@ function gain_ff(gain_value, data) //great! solved clicking! will have to move t
|
||||
}
|
||||
|
||||
function audio_prepare(data) {
|
||||
if (!audio_node) return;
|
||||
var buffer = data;
|
||||
if (audio_compression === "adpcm") {
|
||||
//resampling & ADPCM
|
||||
buffer = audio_codec.decode(buffer);
|
||||
}
|
||||
audio_buffers.push(audio_resampler.process(gain_ff(volume, sdrjs.ConvertI16_F(buffer))));
|
||||
buffer = audio_resampler.process(gain_ff(volume, sdrjs.ConvertI16_F(buffer)));
|
||||
if (audio_node.port) {
|
||||
// AudioWorklets supported
|
||||
audio_node.port.postMessage(buffer, [buffer.buffer]);
|
||||
} else {
|
||||
audio_buffers.push(buffer);
|
||||
}
|
||||
}
|
||||
|
||||
if (!AudioBuffer.prototype.copyToChannel) { //Chrome 36 does not have it, Firefox does
|
||||
@ -1595,13 +1602,13 @@ function audio_buffers_total_length() {
|
||||
return audio_buffers.map(function(b){ return b.length; }).reduce(function(a, b){ return a + b; }, 0);
|
||||
}
|
||||
|
||||
function audio_buffer_progressbar_update() {
|
||||
function audio_buffer_progressbar_update(reportedValue) {
|
||||
if (audio_buffer_progressbar_update_disabled) return;
|
||||
var audio_buffer_value = audio_buffers_total_length() / audio_context.sampleRate;
|
||||
var audio_buffer_value = (reportedValue || audio_buffers_total_length()) / audio_context.sampleRate;
|
||||
audio_buffer_total_average_level_length++;
|
||||
audio_buffer_total_average_level = (audio_buffer_total_average_level * ((audio_buffer_total_average_level_length - 1) / audio_buffer_total_average_level_length)) + (audio_buffer_value / audio_buffer_total_average_level_length);
|
||||
var overrun = audio_buffer_value > audio_buffer_maximal_length_sec;
|
||||
var underrun = audio_buffers.length === 0;
|
||||
var underrun = audio_buffer_value === 0;
|
||||
var text = "buffer";
|
||||
if (overrun) {
|
||||
text = "overrun";
|
||||
@ -1665,21 +1672,14 @@ function parsehash() {
|
||||
|
||||
function audio_preinit() {
|
||||
try {
|
||||
window.AudioContext = window.AudioContext || window.webkitAudioContext;
|
||||
audio_context = new AudioContext();
|
||||
var ctx = window.AudioContext || window.webkitAudioContext;
|
||||
audio_context = new ctx();
|
||||
}
|
||||
catch (e) {
|
||||
divlog('Your browser does not support Web Audio API, which is required for WebRX to run. Please upgrade to a HTML5 compatible browser.', 1);
|
||||
return;
|
||||
}
|
||||
|
||||
if (audio_context.sampleRate < 44100 * 2)
|
||||
audio_buffer_size = 4096;
|
||||
else if (audio_context.sampleRate >= 44100 * 2 && audio_context.sampleRate < 44100 * 4)
|
||||
audio_buffer_size = 4096 * 2;
|
||||
else if (audio_context.sampleRate > 44100 * 4)
|
||||
audio_buffer_size = 4096 * 4;
|
||||
|
||||
//we send our setup packet
|
||||
// TODO this should be moved to another stage of initialization
|
||||
parsehash();
|
||||
@ -1706,21 +1706,55 @@ function audio_init() {
|
||||
audio_debug_time_last_start = audio_debug_time_start;
|
||||
audio_buffer_current_count_debug = 0;
|
||||
|
||||
if (audio_context.sampleRate < 44100 * 2)
|
||||
audio_buffer_size = 4096;
|
||||
else if (audio_context.sampleRate >= 44100 * 2 && audio_context.sampleRate < 44100 * 4)
|
||||
audio_buffer_size = 4096 * 2;
|
||||
else if (audio_context.sampleRate > 44100 * 4)
|
||||
audio_buffer_size = 4096 * 4;
|
||||
|
||||
//https://github.com/0xfe/experiments/blob/master/www/tone/js/sinewave.js
|
||||
audio_initialized = 1; // only tell on_ws_recv() not to call it again
|
||||
|
||||
//on Chrome v36, createJavaScriptNode has been replaced by createScriptProcessor
|
||||
var createjsnode_function = (audio_context.createJavaScriptNode === undefined) ? audio_context.createScriptProcessor.bind(audio_context) : audio_context.createJavaScriptNode.bind(audio_context);
|
||||
audio_node = createjsnode_function(audio_buffer_size, 0, 1);
|
||||
audio_node.onaudioprocess = audio_onprocess;
|
||||
audio_node.connect(audio_context.destination);
|
||||
if (audio_context.audioWorklet) {
|
||||
audio_context.audioWorklet.addModule('static/lib/AudioProcessor.js').then(function(){
|
||||
audio_node = new AudioWorkletNode(audio_context, 'openwebrx-audio-processor', {
|
||||
numberOfInputs: 0,
|
||||
numberOfOutputs: 1,
|
||||
outputChannelCount: [1],
|
||||
processorOptions: {
|
||||
maxLength: audio_buffer_maximal_length_sec,
|
||||
reduceToLength: audio_buffer_decrease_to_on_overrun_sec
|
||||
}
|
||||
});
|
||||
audio_node.connect(audio_context.destination);
|
||||
window.setInterval(function(){
|
||||
audio_node.port.postMessage(JSON.stringify({cmd:'getBuffers'}));
|
||||
}, audio_flush_interval_ms);
|
||||
audio_node.port.addEventListener('message', function(m){
|
||||
var json = JSON.parse(m.data);
|
||||
if (json.buffersize) {
|
||||
audio_buffer_progressbar_update_disabled = false;
|
||||
audio_buffer_progressbar_update(json.buffersize);
|
||||
}
|
||||
});
|
||||
audio_node.port.start();
|
||||
});
|
||||
} else {
|
||||
//on Chrome v36, createJavaScriptNode has been replaced by createScriptProcessor
|
||||
var createjsnode_function = (audio_context.createJavaScriptNode === undefined) ? audio_context.createScriptProcessor.bind(audio_context) : audio_context.createJavaScriptNode.bind(audio_context);
|
||||
audio_node = createjsnode_function(audio_buffer_size, 0, 1);
|
||||
audio_node.onaudioprocess = audio_onprocess;
|
||||
audio_node.connect(audio_context.destination);
|
||||
window.setInterval(audio_flush, audio_flush_interval_ms);
|
||||
}
|
||||
|
||||
// --- Resampling ---
|
||||
//https://github.com/grantgalitz/XAudioJS/blob/master/XAudioServer.js
|
||||
//audio_resampler = new Resampler(audio_received_sample_rate, audio_context.sampleRate, 1, audio_buffer_size, true);
|
||||
//audio_input_buffer_size = audio_buffer_size*(audio_received_sample_rate/audio_context.sampleRate);
|
||||
webrx_set_param("audio_rate", audio_context.sampleRate);
|
||||
|
||||
window.setInterval(audio_flush, audio_flush_interval_ms);
|
||||
divlog('Web Audio API succesfully initialized, sample rate: ' + audio_context.sampleRate.toString() + " sps");
|
||||
initialize_demodulator();
|
||||
|
||||
@ -2224,7 +2258,8 @@ function debug_audio() {
|
||||
var audio_min_rate = audio_context.sampleRate * .25;
|
||||
progressbar_set(e("openwebrx-bar-audio-output"), audio_output_value / audio_max_rate, "Audio output [" + (audio_output_value / 1000).toFixed(1) + " ksps]", audio_output_value > audio_max_rate || audio_output_value < audio_min_rate);
|
||||
|
||||
audio_buffer_progressbar_update();
|
||||
// disable when audioworklets used
|
||||
if (!audio_node.port) audio_buffer_progressbar_update();
|
||||
|
||||
var debug_ws_time_taken = (time_now - debug_ws_time_start) / 1000;
|
||||
var network_speed_value = debug_ws_data_received / debug_ws_time_taken;
|
||||
|
Loading…
Reference in New Issue
Block a user