Reformat dom::servoparser

This commit is contained in:
Anthony Ramine 2017-01-25 13:49:36 +01:00
parent 2b623fcc7a
commit 90839305ae

View file

@ -89,16 +89,11 @@ enum LastChunkState {
} }
impl ServoParser { impl ServoParser {
pub fn parse_html_document( pub fn parse_html_document(document: &Document, input: DOMString, url: ServoUrl, owner: Option<PipelineId>) {
document: &Document, let parser = ServoParser::new(document,
input: DOMString, owner,
url: ServoUrl, Tokenizer::Html(self::html::Tokenizer::new(document, url, None)),
owner: Option<PipelineId>) { LastChunkState::NotReceived);
let parser = ServoParser::new(
document,
owner,
Tokenizer::Html(self::html::Tokenizer::new(document, url, None)),
LastChunkState::NotReceived);
parser.parse_chunk(String::from(input)); parser.parse_chunk(String::from(input));
} }
@ -112,48 +107,49 @@ impl ServoParser {
// Step 1. // Step 1.
let loader = DocumentLoader::new_with_threads(context_document.loader().resource_threads().clone(), let loader = DocumentLoader::new_with_threads(context_document.loader().resource_threads().clone(),
Some(url.clone())); Some(url.clone()));
let document = Document::new(window, None, Some(url.clone()), let document = Document::new(window,
None,
Some(url.clone()),
context_document.origin().alias(), context_document.origin().alias(),
IsHTMLDocument::HTMLDocument, IsHTMLDocument::HTMLDocument,
None, None, None,
None,
DocumentSource::FromParser, DocumentSource::FromParser,
loader, loader,
None, None); None,
None);
// Step 2. // Step 2.
document.set_quirks_mode(context_document.quirks_mode()); document.set_quirks_mode(context_document.quirks_mode());
// Step 11. // Step 11.
let form = context_node.inclusive_ancestors() let form = context_node.inclusive_ancestors()
.find(|element| element.is::<HTMLFormElement>()); .find(|element| element.is::<HTMLFormElement>());
let fragment_context = FragmentContext { let fragment_context = FragmentContext {
context_elem: context_node, context_elem: context_node,
form_elem: form.r(), form_elem: form.r(),
}; };
let parser = ServoParser::new( let parser = ServoParser::new(&document,
&document, None,
None, Tokenizer::Html(self::html::Tokenizer::new(&document,
Tokenizer::Html( url.clone(),
self::html::Tokenizer::new(&document, url.clone(), Some(fragment_context))), Some(fragment_context))),
LastChunkState::Received); LastChunkState::Received);
parser.parse_chunk(String::from(input)); parser.parse_chunk(String::from(input));
// Step 14. // Step 14.
let root_element = document.GetDocumentElement().expect("no document element"); let root_element = document.GetDocumentElement().expect("no document element");
FragmentParsingResult { inner: root_element.upcast::<Node>().children() } FragmentParsingResult {
inner: root_element.upcast::<Node>().children(),
}
} }
pub fn parse_xml_document( pub fn parse_xml_document(document: &Document, input: DOMString, url: ServoUrl, owner: Option<PipelineId>) {
document: &Document, let parser = ServoParser::new(document,
input: DOMString, owner,
url: ServoUrl, Tokenizer::Xml(self::xml::Tokenizer::new(document, url)),
owner: Option<PipelineId>) { LastChunkState::NotReceived);
let parser = ServoParser::new(
document,
owner,
Tokenizer::Xml(self::xml::Tokenizer::new(document, url)),
LastChunkState::NotReceived);
parser.parse_chunk(String::from(input)); parser.parse_chunk(String::from(input));
} }
@ -179,7 +175,8 @@ impl ServoParser {
assert!(self.suspended.get()); assert!(self.suspended.get());
self.suspended.set(false); self.suspended.set(false);
mem::swap(&mut *self.script_input.borrow_mut(), &mut *self.network_input.borrow_mut()); mem::swap(&mut *self.script_input.borrow_mut(),
&mut *self.network_input.borrow_mut());
while let Some(chunk) = self.script_input.borrow_mut().pop_front() { while let Some(chunk) = self.script_input.borrow_mut().pop_front() {
self.network_input.borrow_mut().push_back(chunk); self.network_input.borrow_mut().push_back(chunk);
} }
@ -236,12 +233,11 @@ impl ServoParser {
} }
#[allow(unrooted_must_root)] #[allow(unrooted_must_root)]
fn new_inherited( fn new_inherited(document: &Document,
document: &Document, pipeline: Option<PipelineId>,
pipeline: Option<PipelineId>, tokenizer: Tokenizer,
tokenizer: Tokenizer, last_chunk_state: LastChunkState)
last_chunk_state: LastChunkState) -> Self {
-> Self {
ServoParser { ServoParser {
reflector: Reflector::new(), reflector: Reflector::new(),
document: JS::from_ref(document), document: JS::from_ref(document),
@ -256,16 +252,14 @@ impl ServoParser {
} }
#[allow(unrooted_must_root)] #[allow(unrooted_must_root)]
fn new( fn new(document: &Document,
document: &Document, pipeline: Option<PipelineId>,
pipeline: Option<PipelineId>, tokenizer: Tokenizer,
tokenizer: Tokenizer, last_chunk_state: LastChunkState)
last_chunk_state: LastChunkState) -> Root<Self> {
-> Root<Self> { reflect_dom_object(box ServoParser::new_inherited(document, pipeline, tokenizer, last_chunk_state),
reflect_dom_object( document.window(),
box ServoParser::new_inherited(document, pipeline, tokenizer, last_chunk_state), ServoParserBinding::Wrap)
document.window(),
ServoParserBinding::Wrap)
} }
fn push_input_chunk(&self, chunk: String) { fn push_input_chunk(&self, chunk: String) {
@ -313,7 +307,7 @@ impl ServoParser {
} }
fn tokenize<F>(&self, mut feed: F) fn tokenize<F>(&self, mut feed: F)
where F: FnMut(&mut Tokenizer) -> Result<(), Root<HTMLScriptElement>> where F: FnMut(&mut Tokenizer) -> Result<(), Root<HTMLScriptElement>>,
{ {
loop { loop {
assert!(!self.suspended.get()); assert!(!self.suspended.get());
@ -356,7 +350,9 @@ impl ServoParser {
self.document.disarm_reflow_timeout(); self.document.disarm_reflow_timeout();
self.document.upcast::<Node>().dirty(NodeDamage::OtherNodeDamage); self.document.upcast::<Node>().dirty(NodeDamage::OtherNodeDamage);
let window = self.document.window(); let window = self.document.window();
window.reflow(ReflowGoal::ForDisplay, ReflowQueryType::NoQuery, ReflowReason::FirstLoad); window.reflow(ReflowGoal::ForDisplay,
ReflowQueryType::NoQuery,
ReflowReason::FirstLoad);
} }
// Steps 3-12 are in another castle, namely finish_load. // Steps 3-12 are in another castle, namely finish_load.
@ -455,14 +451,13 @@ impl FetchResponseListener for ParserContext {
fn process_request_eof(&mut self) {} fn process_request_eof(&mut self) {}
fn process_response(&mut self, fn process_response(&mut self, meta_result: Result<FetchMetadata, NetworkError>) {
meta_result: Result<FetchMetadata, NetworkError>) {
let mut ssl_error = None; let mut ssl_error = None;
let metadata = match meta_result { let metadata = match meta_result {
Ok(meta) => { Ok(meta) => {
Some(match meta { Some(match meta {
FetchMetadata::Unfiltered(m) => m, FetchMetadata::Unfiltered(m) => m,
FetchMetadata::Filtered { unsafe_, .. } => unsafe_ FetchMetadata::Filtered { unsafe_, .. } => unsafe_,
}) })
}, },
Err(NetworkError::SslValidation(url, reason)) => { Err(NetworkError::SslValidation(url, reason)) => {
@ -474,10 +469,8 @@ impl FetchResponseListener for ParserContext {
}, },
Err(_) => None, Err(_) => None,
}; };
let content_type = let content_type = metadata.clone().and_then(|meta| meta.content_type).map(Serde::into_inner);
metadata.clone().and_then(|meta| meta.content_type).map(Serde::into_inner); let parser = match ScriptThread::page_headers_available(&self.id, metadata) {
let parser = match ScriptThread::page_headers_available(&self.id,
metadata) {
Some(parser) => parser, Some(parser) => parser,
None => return, None => return,
}; };
@ -505,7 +498,8 @@ impl FetchResponseListener for ParserContext {
parser.parse_sync(); parser.parse_sync();
parser.tokenizer.borrow_mut().set_plaintext_state(); parser.tokenizer.borrow_mut().set_plaintext_state();
}, },
Some(ContentType(Mime(TopLevel::Text, SubLevel::Html, _))) => { // Handle text/html Some(ContentType(Mime(TopLevel::Text, SubLevel::Html, _))) => {
// Handle text/html
if let Some(reason) = ssl_error { if let Some(reason) = ssl_error {
self.is_synthesized_document = true; self.is_synthesized_document = true;
let page_bytes = read_resource_file("badcert.html").unwrap(); let page_bytes = read_resource_file("badcert.html").unwrap();
@ -524,7 +518,8 @@ impl FetchResponseListener for ParserContext {
// Show warning page for unknown mime types. // Show warning page for unknown mime types.
let page = format!("<html><body><p>Unknown content type ({}/{}).</p></body></html>", let page = format!("<html><body><p>Unknown content type ({}/{}).</p></body></html>",
toplevel.as_str(), sublevel.as_str()); toplevel.as_str(),
sublevel.as_str());
self.is_synthesized_document = true; self.is_synthesized_document = true;
parser.push_input_chunk(page); parser.push_input_chunk(page);
parser.parse_sync(); parser.parse_sync();
@ -532,7 +527,7 @@ impl FetchResponseListener for ParserContext {
None => { None => {
// No content-type header. // No content-type header.
// Merge with #4212 when fixed. // Merge with #4212 when fixed.
} },
} }
} }
@ -595,6 +590,6 @@ fn insert(parent: &Node, reference_child: Option<&Node>, child: NodeOrText<JS<No
let text = Text::new(String::from(t).into(), &parent.owner_doc()); let text = Text::new(String::from(t).into(), &parent.owner_doc());
parent.InsertBefore(text.upcast(), reference_child).unwrap(); parent.InsertBefore(text.upcast(), reference_child).unwrap();
} }
} },
} }
} }