Properly implement audio context state changes

This commit is contained in:
Fernando Jiménez Moreno 2018-06-26 11:51:57 +02:00
parent 9f15bdd6cc
commit aed57252b1
4 changed files with 214 additions and 56 deletions

View file

@ -11,14 +11,14 @@ use dom::bindings::codegen::Bindings::BaseAudioContextBinding::BaseAudioContextB
use dom::bindings::error::{Error, Fallible}; use dom::bindings::error::{Error, Fallible};
use dom::bindings::inheritance::Castable; use dom::bindings::inheritance::Castable;
use dom::bindings::num::Finite; use dom::bindings::num::Finite;
use dom::bindings::refcounted::Trusted; use dom::bindings::refcounted::{Trusted, TrustedPromise};
use dom::bindings::reflector::{DomObject, reflect_dom_object}; use dom::bindings::reflector::{DomObject, reflect_dom_object};
use dom::bindings::root::DomRoot; use dom::bindings::root::DomRoot;
use dom::globalscope::GlobalScope; use dom::globalscope::GlobalScope;
use dom::promise::Promise; use dom::promise::Promise;
use dom::window::Window; use dom::window::Window;
use dom_struct::dom_struct; use dom_struct::dom_struct;
use servo_media::audio::graph::{LatencyCategory, RealTimeAudioGraphOptions}; use servo_media::audio::context::{LatencyCategory, ProcessingState, RealTimeAudioContextOptions};
use std::rc::Rc; use std::rc::Rc;
use task_source::TaskSource; use task_source::TaskSource;
@ -99,14 +99,13 @@ impl AudioContextMethods for AudioContext {
let promise = Promise::new(&self.global()); let promise = Promise::new(&self.global());
// Step 2. // Step 2.
let state = self.context.State(); if self.context.control_thread_state() == ProcessingState::Closed {
if state == AudioContextState::Closed {
promise.reject_error(Error::InvalidState); promise.reject_error(Error::InvalidState);
return promise; return promise;
} }
// Step 3. // Step 3.
if state == AudioContextState::Suspended { if self.context.State() == AudioContextState::Suspended {
promise.resolve_native(&()); promise.resolve_native(&());
return promise; return promise;
} }
@ -114,18 +113,36 @@ impl AudioContextMethods for AudioContext {
// Steps 4 and 5. // Steps 4 and 5.
let window = DomRoot::downcast::<Window>(self.global()).unwrap(); let window = DomRoot::downcast::<Window>(self.global()).unwrap();
let task_source = window.dom_manipulation_task_source(); let task_source = window.dom_manipulation_task_source();
let trusted_promise = TrustedPromise::new(promise.clone());
let this = Trusted::new(self); match self.context.audio_context_impl().suspend() {
task_source.queue(task!(suspend: move || { Ok(_) => {
let this = this.root(); let base_context = Trusted::new(&self.context);
this.context.audio_graph().suspend(); let context = Trusted::new(self);
}), window.upcast()).unwrap(); let _ = task_source.queue(task!(suspend_ok: move || {
let base_context = base_context.root();
task_source.queue_simple_event( let context = context.root();
self.upcast(), let promise = trusted_promise.root();
atom!("statechange"), promise.resolve_native(&());
&window if base_context.State() != AudioContextState::Suspended {
); base_context.set_state_attribute(AudioContextState::Suspended);
let window = DomRoot::downcast::<Window>(context.global()).unwrap();
window.dom_manipulation_task_source().queue_simple_event(
context.upcast(),
atom!("statechange"),
&window
);
}
}), window.upcast());
},
Err(_) => {
// The spec does not define the error case and `suspend` should
// never fail, but we handle the case here for completion.
let _ = task_source.queue(task!(suspend_error: move || {
let promise = trusted_promise.root();
promise.reject_error(Error::Type("Something went wrong".to_owned()));
}), window.upcast());
},
};
// Step 6. // Step 6.
promise promise
@ -133,8 +150,58 @@ impl AudioContextMethods for AudioContext {
#[allow(unrooted_must_root)] #[allow(unrooted_must_root)]
fn Close(&self) -> Rc<Promise> { fn Close(&self) -> Rc<Promise> {
// TODO // Step 1.
Promise::new(&self.global()) let promise = Promise::new(&self.global());
// Step 2.
if self.context.control_thread_state() == ProcessingState::Closed {
promise.reject_error(Error::InvalidState);
return promise;
}
// Step 3.
if self.context.State() == AudioContextState::Closed {
promise.resolve_native(&());
return promise;
}
// Steps 4 and 5.
let window = DomRoot::downcast::<Window>(self.global()).unwrap();
let task_source = window.dom_manipulation_task_source();
let trusted_promise = TrustedPromise::new(promise.clone());
match self.context.audio_context_impl().close() {
Ok(_) => {
let base_context = Trusted::new(&self.context);
let context = Trusted::new(self);
let _ = task_source.queue(task!(suspend_ok: move || {
let base_context = base_context.root();
let context = context.root();
let promise = trusted_promise.root();
promise.resolve_native(&());
if base_context.State() != AudioContextState::Closed {
base_context.set_state_attribute(AudioContextState::Closed);
let window = DomRoot::downcast::<Window>(context.global()).unwrap();
window.dom_manipulation_task_source().queue_simple_event(
context.upcast(),
atom!("statechange"),
&window
);
}
}), window.upcast());
},
Err(_) => {
// The spec does not define the error case and `suspend` should
// never fail, but we handle the case here for completion.
let _ = task_source.queue(task!(suspend_error: move || {
let promise = trusted_promise.root();
promise.reject_error(Error::Type("Something went wrong".to_owned()));
}), window.upcast());
},
};
// Step 6.
promise
} }
} }
@ -148,7 +215,7 @@ impl From<AudioContextLatencyCategory> for LatencyCategory {
} }
} }
impl<'a> From<&'a AudioContextOptions> for RealTimeAudioGraphOptions { impl<'a> From<&'a AudioContextOptions> for RealTimeAudioContextOptions {
fn from(options: &AudioContextOptions) -> Self { fn from(options: &AudioContextOptions) -> Self {
Self { Self {
sample_rate: *options.sampleRate.unwrap_or(Finite::wrap(48000.)), sample_rate: *options.sampleRate.unwrap_or(Finite::wrap(48000.)),

View file

@ -10,7 +10,7 @@ use dom::bindings::reflector::Reflector;
use dom::bindings::root::DomRoot; use dom::bindings::root::DomRoot;
use dom::audioparam::AudioParam; use dom::audioparam::AudioParam;
use dom_struct::dom_struct; use dom_struct::dom_struct;
use servo_media::audio::graph_impl::NodeId; use servo_media::audio::graph::NodeId;
use servo_media::audio::node::AudioNodeType; use servo_media::audio::node::AudioNodeType;
use std::cell::Cell; use std::cell::Cell;

View file

@ -9,7 +9,7 @@ use dom::bindings::codegen::Bindings::AudioNodeBinding::{ChannelCountMode, Chann
use dom::bindings::codegen::Bindings::BaseAudioContextBinding::BaseAudioContextMethods; use dom::bindings::codegen::Bindings::BaseAudioContextBinding::BaseAudioContextMethods;
use dom::bindings::codegen::Bindings::BaseAudioContextBinding::AudioContextState; use dom::bindings::codegen::Bindings::BaseAudioContextBinding::AudioContextState;
use dom::bindings::codegen::Bindings::OscillatorNodeBinding::OscillatorOptions; use dom::bindings::codegen::Bindings::OscillatorNodeBinding::OscillatorOptions;
use dom::bindings::error::Error; use dom::bindings::error::{Error, ErrorResult};
use dom::bindings::inheritance::Castable; use dom::bindings::inheritance::Castable;
use dom::bindings::num::Finite; use dom::bindings::num::Finite;
use dom::bindings::refcounted::Trusted; use dom::bindings::refcounted::Trusted;
@ -21,30 +21,42 @@ use dom::promise::Promise;
use dom::window::Window; use dom::window::Window;
use dom_struct::dom_struct; use dom_struct::dom_struct;
use servo_media::ServoMedia; use servo_media::ServoMedia;
use servo_media::audio::graph::{AudioGraph, ProcessingState}; use servo_media::audio::context::{AudioContext, ProcessingState};
use servo_media::audio::graph::{OfflineAudioGraphOptions, RealTimeAudioGraphOptions}; use servo_media::audio::context::{OfflineAudioContextOptions, RealTimeAudioContextOptions};
use servo_media::audio::graph_impl::NodeId; use servo_media::audio::graph::NodeId;
use servo_media::audio::node::AudioNodeType; use servo_media::audio::node::AudioNodeType;
use std::cell::Cell;
use std::collections::VecDeque;
use std::mem;
use std::rc::Rc; use std::rc::Rc;
use task_source::TaskSource; use task_source::TaskSource;
pub enum BaseAudioContextOptions { pub enum BaseAudioContextOptions {
AudioContext(RealTimeAudioGraphOptions), AudioContext(RealTimeAudioContextOptions),
OfflineAudioContext(OfflineAudioGraphOptions), OfflineAudioContext(OfflineAudioContextOptions),
} }
#[dom_struct] #[dom_struct]
pub struct BaseAudioContext { pub struct BaseAudioContext {
reflector_: Reflector, reflector_: Reflector,
#[ignore_malloc_size_of = "servo_media"] #[ignore_malloc_size_of = "servo_media"]
audio_graph: AudioGraph, audio_context_impl: AudioContext,
/// https://webaudio.github.io/web-audio-api/#dom-baseaudiocontext-destination /// https://webaudio.github.io/web-audio-api/#dom-baseaudiocontext-destination
destination: Option<DomRoot<AudioDestinationNode>>, destination: Option<DomRoot<AudioDestinationNode>>,
/// https://webaudio.github.io/web-audio-api/#dom-baseaudiocontext-samplerate /// Resume promises which are soon to be fulfilled by a queued task.
sample_rate: f32, #[ignore_malloc_size_of = "promises are hard"]
in_flight_resume_promises_queue: DomRefCell<VecDeque<(Box<[Rc<Promise>]>, ErrorResult)>>,
/// https://webaudio.github.io/web-audio-api/#pendingresumepromises /// https://webaudio.github.io/web-audio-api/#pendingresumepromises
#[ignore_malloc_size_of = "promises are hard"] #[ignore_malloc_size_of = "promises are hard"]
pending_resume_promises: DomRefCell<Vec<Rc<Promise>>>, pending_resume_promises: DomRefCell<Vec<Rc<Promise>>>,
/// https://webaudio.github.io/web-audio-api/#dom-baseaudiocontext-samplerate
sample_rate: f32,
/// https://webaudio.github.io/web-audio-api/#dom-baseaudiocontext-state
/// Although servo-media already keeps track of the control thread state,
/// we keep a state flag here as well. This is so that we can synchronously
/// throw when trying to do things on the context when the context has just
/// been "closed()".
state: Cell<AudioContextState>,
} }
impl BaseAudioContext { impl BaseAudioContext {
@ -63,10 +75,12 @@ impl BaseAudioContext {
let mut context = BaseAudioContext { let mut context = BaseAudioContext {
reflector_: Reflector::new(), reflector_: Reflector::new(),
audio_graph: ServoMedia::get().unwrap().create_audio_graph(Some(options.into())), audio_context_impl: ServoMedia::get().unwrap().create_audio_context(options.into()),
destination: None, destination: None,
sample_rate, in_flight_resume_promises_queue: Default::default(),
pending_resume_promises: Default::default(), pending_resume_promises: Default::default(),
sample_rate,
state: Cell::new(AudioContextState::Suspended),
}; };
let mut options = unsafe { AudioNodeOptions::empty(global.get_cx()) }; let mut options = unsafe { AudioNodeOptions::empty(global.get_cx()) };
@ -79,36 +93,110 @@ impl BaseAudioContext {
context context
} }
pub fn audio_graph(&self) -> &AudioGraph { pub fn audio_context_impl(&self) -> &AudioContext {
&self.audio_graph &self.audio_context_impl
} }
pub fn create_node_engine(&self, node_type: AudioNodeType) -> NodeId { pub fn create_node_engine(&self, node_type: AudioNodeType) -> NodeId {
self.audio_graph.create_node(node_type) self.audio_context_impl.create_node(node_type)
} }
// https://webaudio.github.io/web-audio-api/#allowed-to-start // https://webaudio.github.io/web-audio-api/#allowed-to-start
pub fn is_allowed_to_start(&self) -> bool { pub fn is_allowed_to_start(&self) -> bool {
self.audio_graph.state() == ProcessingState::Suspended self.state.get() == AudioContextState::Suspended
}
/// Takes the pending resume promises.
///
/// The result with which these promises will be fulfilled is passed here
/// and this method returns nothing because we actually just move the
/// current list of pending resume promises to the
/// `in_flight_resume_promises_queue` field.
///
/// Each call to this method must be followed by a call to
/// `fulfill_in_flight_resume_promises`, to actually fulfill the promises
/// which were taken and moved to the in-flight queue.
#[allow(unrooted_must_root)]
fn take_pending_resume_promises(&self, result: ErrorResult) {
let pending_resume_promises = mem::replace(
&mut *self.pending_resume_promises.borrow_mut(),
vec![],
);
self.in_flight_resume_promises_queue.borrow_mut().push_back((
pending_resume_promises.into(),
result,
));
}
/// Fulfills the next in-flight resume promises queue after running a closure.
///
/// See the comment on `take_pending_resume_promises` for why this method
/// does not take a list of promises to fulfill. Callers cannot just pop
/// the front list off of `in_flight_resume_promises_queue` and later fulfill
/// the promises because that would mean putting
/// `#[allow(unrooted_must_root)]` on even more functions, potentially
/// hiding actual safety bugs.
#[allow(unrooted_must_root)]
fn fulfill_in_flight_resume_promises<F>(&self, f: F)
where
F: FnOnce(),
{
let (promises, result) = self.in_flight_resume_promises_queue
.borrow_mut()
.pop_front()
.expect("there should be at least one list of in flight resume promises");
f();
for promise in &*promises {
match result {
Ok(ref value) => promise.resolve_native(value),
Err(ref error) => promise.reject_error(error.clone()),
}
}
}
/// Control thread processing state
pub fn control_thread_state(&self) -> ProcessingState {
self.audio_context_impl.state()
}
/// Set audio context state
pub fn set_state_attribute(&self, state: AudioContextState) {
self.state.set(state);
} }
pub fn resume(&self) { pub fn resume(&self) {
let window = DomRoot::downcast::<Window>(self.global()).unwrap(); let window = DomRoot::downcast::<Window>(self.global()).unwrap();
let task_source = window.dom_manipulation_task_source(); let task_source = window.dom_manipulation_task_source();
// Set the state attribute to `running` and start rendering audio.
let this = Trusted::new(self); let this = Trusted::new(self);
task_source.queue(task!(resume: move || {
let this = this.root();
this.audio_graph.resume();
}), window.upcast()).unwrap();
// Queue a task to fire a simple event named `statechange` at the AudioContext. // Set the rendering thread state to 'running' and start
task_source.queue_simple_event( // rendering the audio graph.
self.upcast(), match self.audio_context_impl.resume() {
atom!("statechange"), Ok(()) => {
&window self.take_pending_resume_promises(Ok(()));
); let _ = task_source.queue(task!(resume_success: move || {
let this = this.root();
this.fulfill_in_flight_resume_promises(|| {
if this.state.get() != AudioContextState::Running {
this.state.set(AudioContextState::Running);
let window = DomRoot::downcast::<Window>(this.global()).unwrap();
window.dom_manipulation_task_source().queue_simple_event(
this.upcast(),
atom!("statechange"),
&window
);
}
});
}), window.upcast());
},
Err(()) => {
self.take_pending_resume_promises(Err(Error::Type("Something went wrong".to_owned())));
let _ = task_source.queue(task!(resume_error: move || {
let this = this.root();
this.fulfill_in_flight_resume_promises(|| {});
}), window.upcast());
}
}
} }
} }
@ -120,13 +208,13 @@ impl BaseAudioContextMethods for BaseAudioContext {
// https://webaudio.github.io/web-audio-api/#dom-baseaudiocontext-currenttime // https://webaudio.github.io/web-audio-api/#dom-baseaudiocontext-currenttime
fn CurrentTime(&self) -> Finite<f64> { fn CurrentTime(&self) -> Finite<f64> {
let current_time = self.audio_graph.current_time(); let current_time = self.audio_context_impl.current_time();
Finite::wrap(current_time) Finite::wrap(current_time)
} }
// https://webaudio.github.io/web-audio-api/#dom-baseaudiocontext-state // https://webaudio.github.io/web-audio-api/#dom-baseaudiocontext-state
fn State(&self) -> AudioContextState { fn State(&self) -> AudioContextState {
self.audio_graph.state().into() self.state.get()
} }
// https://webaudio.github.io/web-audio-api/#dom-baseaudiocontext-resume // https://webaudio.github.io/web-audio-api/#dom-baseaudiocontext-resume
@ -136,21 +224,23 @@ impl BaseAudioContextMethods for BaseAudioContext {
let promise = Promise::new(&self.global()); let promise = Promise::new(&self.global());
// Step 2. // Step 2.
let state = self.audio_graph.state(); if self.audio_context_impl.state() == ProcessingState::Closed {
if state == ProcessingState::Closed {
promise.reject_error(Error::InvalidState); promise.reject_error(Error::InvalidState);
return promise; return promise;
} }
// Step 3. // Step 3.
if state == ProcessingState::Running { if self.state.get() == AudioContextState::Running {
promise.resolve_native(&()); promise.resolve_native(&());
return promise; return promise;
} }
// Push the promise into the queue to avoid passing a reference to
// `resume()`. This way we limit the usage of #[allow(unrooted_must_root)]
self.pending_resume_promises.borrow_mut().push(promise.clone());
// Step 4. // Step 4.
if !self.is_allowed_to_start() { if !self.is_allowed_to_start() {
self.pending_resume_promises.borrow_mut().push(promise.clone());
return promise; return promise;
} }

View file

@ -80,8 +80,8 @@ use offscreen_gl_context::GLLimits;
use parking_lot::RwLock; use parking_lot::RwLock;
use profile_traits::mem::ProfilerChan as MemProfilerChan; use profile_traits::mem::ProfilerChan as MemProfilerChan;
use profile_traits::time::ProfilerChan as TimeProfilerChan; use profile_traits::time::ProfilerChan as TimeProfilerChan;
use servo_media::audio::graph::AudioGraph; use servo_media::audio::context::AudioContext;
use servo_media::audio::graph_impl::NodeId; use servo_media::audio::graph::NodeId;
use script_layout_interface::OpaqueStyleAndLayoutData; use script_layout_interface::OpaqueStyleAndLayoutData;
use script_layout_interface::reporter::CSSErrorReporter; use script_layout_interface::reporter::CSSErrorReporter;
use script_layout_interface::rpc::LayoutRPC; use script_layout_interface::rpc::LayoutRPC;
@ -432,6 +432,7 @@ unsafe_no_jsmanaged_fields!(InteractiveWindow);
unsafe_no_jsmanaged_fields!(CanvasId); unsafe_no_jsmanaged_fields!(CanvasId);
unsafe_no_jsmanaged_fields!(SourceSet); unsafe_no_jsmanaged_fields!(SourceSet);
unsafe_no_jsmanaged_fields!(AudioGraph); unsafe_no_jsmanaged_fields!(AudioGraph);
unsafe_no_jsmanaged_fields!(AudioContext);
unsafe_no_jsmanaged_fields!(NodeId); unsafe_no_jsmanaged_fields!(NodeId);
unsafe impl<'a> JSTraceable for &'a str { unsafe impl<'a> JSTraceable for &'a str {