ladybird/Libraries/LibWeb/WebAudio/OfflineAudioContext.cpp
Ben Eidson 01947ded23 LibWeb/WebAudio: Implement basic startRendering
Adds passing WPT. Does not handle actually rendering audio yet.
2025-10-26 14:19:21 +01:00

197 lines
9.3 KiB
C++
Raw Blame History

This file contains ambiguous Unicode characters

This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

/*
* Copyright (c) 2024, Shannon Booth <shannon@serenityos.org>
* Copyright (c) 2025, Ben Eidson <b.e.eidson@gmail.com>
*
* SPDX-License-Identifier: BSD-2-Clause
*/
#include <LibWeb/Bindings/Intrinsics.h>
#include <LibWeb/DOM/Event.h>
#include <LibWeb/HTML/EventNames.h>
#include <LibWeb/HTML/Navigable.h>
#include <LibWeb/HTML/Scripting/TemporaryExecutionContext.h>
#include <LibWeb/HTML/Window.h>
#include <LibWeb/WebAudio/AudioBuffer.h>
#include <LibWeb/WebAudio/AudioDestinationNode.h>
#include <LibWeb/WebAudio/OfflineAudioContext.h>
namespace Web::WebAudio {
GC_DEFINE_ALLOCATOR(OfflineAudioContext);
// https://webaudio.github.io/web-audio-api/#dom-offlineaudiocontext-offlineaudiocontext
WebIDL::ExceptionOr<GC::Ref<OfflineAudioContext>> OfflineAudioContext::construct_impl(JS::Realm& realm, OfflineAudioContextOptions const& context_options)
{
// AD-HOC: This spec text is currently only mentioned in the constructor overload that takes separate arguments,
// but these parameters should be validated for both constructors.
// A NotSupportedError exception MUST be thrown if any of the arguments is negative, zero, or outside its nominal range.
TRY(verify_audio_options_inside_nominal_range(realm, context_options.number_of_channels, context_options.length, context_options.sample_rate));
// Let c be a new OfflineAudioContext object. Initialize c as follows:
auto c = realm.create<OfflineAudioContext>(realm, context_options.number_of_channels, context_options.length, context_options.sample_rate);
// 1. Set the [[control thread state]] for c to "suspended".
c->set_control_state(Bindings::AudioContextState::Suspended);
// 2. Set the [[rendering thread state]] for c to "suspended".
c->set_rendering_state(Bindings::AudioContextState::Suspended);
// FIXME: 3. Determine the [[render quantum size]] for this OfflineAudioContext, based on the value of the renderSizeHint:
// 4. Construct an AudioDestinationNode with its channelCount set to contextOptions.numberOfChannels.
c->m_destination = TRY(AudioDestinationNode::construct_impl(realm, c, context_options.number_of_channels));
// FIXME: 5. Let messageChannel be a new MessageChannel.
// FIXME: 6. Let controlSidePort be the value of messageChannels port1 attribute.
// FIXME: 7. Let renderingSidePort be the value of messageChannels port2 attribute.
// FIXME: 8. Let serializedRenderingSidePort be the result of StructuredSerializeWithTransfer(renderingSidePort, « renderingSidePort »).
// FIXME: 9. Set this audioWorklet's port to controlSidePort.
// FIXME: 10. Queue a control message to set the MessagePort on the AudioContextGlobalScope, with serializedRenderingSidePort.
return c;
}
// https://webaudio.github.io/web-audio-api/#dom-offlineaudiocontext-offlineaudiocontext-numberofchannels-length-samplerate
WebIDL::ExceptionOr<GC::Ref<OfflineAudioContext>> OfflineAudioContext::construct_impl(JS::Realm& realm,
WebIDL::UnsignedLong number_of_channels, WebIDL::UnsignedLong length, float sample_rate)
{
return construct_impl(realm, { number_of_channels, length, sample_rate });
}
OfflineAudioContext::~OfflineAudioContext() = default;
// https://webaudio.github.io/web-audio-api/#dom-offlineaudiocontext-startrendering
WebIDL::ExceptionOr<GC::Ref<WebIDL::Promise>> OfflineAudioContext::start_rendering()
{
auto& realm = this->realm();
// 1. If thiss relevant global objects associated Document is not fully active then return a promise rejected with "InvalidStateError" DOMException.
auto& window = as<HTML::Window>(HTML::relevant_global_object(*this));
auto const& associated_document = window.associated_document();
if (!associated_document.is_fully_active()) {
auto error = WebIDL::InvalidStateError::create(realm, "Document is not fully active"_utf16);
return WebIDL::create_rejected_promise_from_exception(realm, error);
}
// AD-HOC: Not in spec explicitly, but this should account for detached iframes too. See /the-offlineaudiocontext-interface/startrendering-after-discard.html WPT.
auto navigable = window.navigable();
if (navigable && navigable->has_been_destroyed()) {
auto error = WebIDL::InvalidStateError::create(realm, "The iframe has been detached"_utf16);
return WebIDL::create_rejected_promise_from_exception(realm, error);
}
// 2. If the [[rendering started]] slot on the OfflineAudioContext is true, return a rejected promise with InvalidStateError, and abort these steps.
if (m_rendering_started) {
auto error = WebIDL::InvalidStateError::create(realm, "Rendering is already started"_utf16);
return WebIDL::create_rejected_promise_from_exception(realm, error);
}
// 3. Set the [[rendering started]] slot of the OfflineAudioContext to true.
m_rendering_started = true;
// 4. Let promise be a new promise.
auto promise = WebIDL::create_promise(realm);
// 5. Create a new AudioBuffer, with a number of channels, length and sample rate equal respectively to the
// numberOfChannels, length and sampleRate values passed to this instances constructor in the contextOptions
// parameter.
auto buffer_result = create_buffer(m_number_of_channels, length(), sample_rate());
// 6. If an exception was thrown during the preceding AudioBuffer constructor call, reject promise with this exception.
if (buffer_result.is_exception()) {
return WebIDL::create_rejected_promise_from_exception(realm, buffer_result.exception());
}
// Assign this buffer to an internal slot [[rendered buffer]] in the OfflineAudioContext.
m_rendered_buffer = buffer_result.release_value();
// 7. Otherwise, in the case that the buffer was successfully constructed, begin offline rendering.
begin_offline_rendering(promise);
// 8. Append promise to [[pending promises]].
m_pending_promises.append(promise);
// 9. Return promise.
return promise;
}
void OfflineAudioContext::begin_offline_rendering(GC::Ref<WebIDL::Promise> promise)
{
auto& realm = this->realm();
// To begin offline rendering, the following steps MUST happen on a rendering thread that is created for the occasion.
// FIXME: 1: Given the current connections and scheduled changes, start rendering length sample-frames of audio into [[rendered buffer]]
// FIXME: 2: For every render quantum, check and suspend rendering if necessary.
// FIXME: 3: If a suspended context is resumed, continue to render the buffer.
// 4: Once the rendering is complete, queue a media element task to execute the following steps:
queue_a_media_element_task(GC::create_function(heap(), [&realm, promise, this]() {
HTML::TemporaryExecutionContext context(realm, HTML::TemporaryExecutionContext::CallbacksEnabled::Yes);
// 4.1 Resolve the promise created by startRendering() with [[rendered buffer]].
WebIDL::resolve_promise(realm, promise, this->m_rendered_buffer);
// AD-HOC: Remove resolved promise from [[pending promises]]
// https://github.com/WebAudio/web-audio-api/issues/2648
m_pending_promises.remove_all_matching([promise](GC::Ref<WebIDL::Promise> const& p) {
return p.ptr() == promise.ptr();
});
// 4.2: Queue a media element task to fire an event named complete at the OfflineAudioContext using OfflineAudioCompletionEvent
// whose renderedBuffer property is set to [[rendered buffer]].
// FIXME: Need to implement OfflineAudioCompletionEvent.
queue_a_media_element_task(GC::create_function(heap(), [&realm, this]() {
this->dispatch_event(DOM::Event::create(realm, HTML::EventNames::complete));
}));
}));
}
WebIDL::ExceptionOr<GC::Ref<WebIDL::Promise>> OfflineAudioContext::resume()
{
return WebIDL::NotSupportedError::create(realm(), "FIXME: Implement OfflineAudioContext::resume"_utf16);
}
WebIDL::ExceptionOr<GC::Ref<WebIDL::Promise>> OfflineAudioContext::suspend(double suspend_time)
{
(void)suspend_time;
return WebIDL::NotSupportedError::create(realm(), "FIXME: Implement OfflineAudioContext::suspend"_utf16);
}
// https://webaudio.github.io/web-audio-api/#dom-offlineaudiocontext-length
WebIDL::UnsignedLong OfflineAudioContext::length() const
{
// The size of the buffer in sample-frames. This is the same as the value of the length parameter for the constructor.
return m_length;
}
// https://webaudio.github.io/web-audio-api/#dom-offlineaudiocontext-oncomplete
GC::Ptr<WebIDL::CallbackType> OfflineAudioContext::oncomplete()
{
return event_handler_attribute(HTML::EventNames::complete);
}
// https://webaudio.github.io/web-audio-api/#dom-offlineaudiocontext-oncomplete
void OfflineAudioContext::set_oncomplete(GC::Ptr<WebIDL::CallbackType> value)
{
set_event_handler_attribute(HTML::EventNames::complete, value);
}
OfflineAudioContext::OfflineAudioContext(JS::Realm& realm, WebIDL::UnsignedLong number_of_channels, WebIDL::UnsignedLong length, float sample_rate)
: BaseAudioContext(realm, sample_rate)
, m_length(length)
, m_number_of_channels(number_of_channels)
{
}
void OfflineAudioContext::initialize(JS::Realm& realm)
{
WEB_SET_PROTOTYPE_FOR_INTERFACE(OfflineAudioContext);
Base::initialize(realm);
}
void OfflineAudioContext::visit_edges(Cell::Visitor& visitor)
{
Base::visit_edges(visitor);
visitor.visit(m_rendered_buffer);
}
}