mirror of
https://github.com/servo/servo.git
synced 2025-08-05 13:40:08 +01:00
Reformat dom::servoparser
This commit is contained in:
parent
2b623fcc7a
commit
90839305ae
1 changed files with 56 additions and 61 deletions
|
@ -89,16 +89,11 @@ enum LastChunkState {
|
||||||
}
|
}
|
||||||
|
|
||||||
impl ServoParser {
|
impl ServoParser {
|
||||||
pub fn parse_html_document(
|
pub fn parse_html_document(document: &Document, input: DOMString, url: ServoUrl, owner: Option<PipelineId>) {
|
||||||
document: &Document,
|
let parser = ServoParser::new(document,
|
||||||
input: DOMString,
|
owner,
|
||||||
url: ServoUrl,
|
Tokenizer::Html(self::html::Tokenizer::new(document, url, None)),
|
||||||
owner: Option<PipelineId>) {
|
LastChunkState::NotReceived);
|
||||||
let parser = ServoParser::new(
|
|
||||||
document,
|
|
||||||
owner,
|
|
||||||
Tokenizer::Html(self::html::Tokenizer::new(document, url, None)),
|
|
||||||
LastChunkState::NotReceived);
|
|
||||||
parser.parse_chunk(String::from(input));
|
parser.parse_chunk(String::from(input));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -112,48 +107,49 @@ impl ServoParser {
|
||||||
// Step 1.
|
// Step 1.
|
||||||
let loader = DocumentLoader::new_with_threads(context_document.loader().resource_threads().clone(),
|
let loader = DocumentLoader::new_with_threads(context_document.loader().resource_threads().clone(),
|
||||||
Some(url.clone()));
|
Some(url.clone()));
|
||||||
let document = Document::new(window, None, Some(url.clone()),
|
let document = Document::new(window,
|
||||||
|
None,
|
||||||
|
Some(url.clone()),
|
||||||
context_document.origin().alias(),
|
context_document.origin().alias(),
|
||||||
IsHTMLDocument::HTMLDocument,
|
IsHTMLDocument::HTMLDocument,
|
||||||
None, None,
|
None,
|
||||||
|
None,
|
||||||
DocumentSource::FromParser,
|
DocumentSource::FromParser,
|
||||||
loader,
|
loader,
|
||||||
None, None);
|
None,
|
||||||
|
None);
|
||||||
|
|
||||||
// Step 2.
|
// Step 2.
|
||||||
document.set_quirks_mode(context_document.quirks_mode());
|
document.set_quirks_mode(context_document.quirks_mode());
|
||||||
|
|
||||||
// Step 11.
|
// Step 11.
|
||||||
let form = context_node.inclusive_ancestors()
|
let form = context_node.inclusive_ancestors()
|
||||||
.find(|element| element.is::<HTMLFormElement>());
|
.find(|element| element.is::<HTMLFormElement>());
|
||||||
let fragment_context = FragmentContext {
|
let fragment_context = FragmentContext {
|
||||||
context_elem: context_node,
|
context_elem: context_node,
|
||||||
form_elem: form.r(),
|
form_elem: form.r(),
|
||||||
};
|
};
|
||||||
|
|
||||||
let parser = ServoParser::new(
|
let parser = ServoParser::new(&document,
|
||||||
&document,
|
None,
|
||||||
None,
|
Tokenizer::Html(self::html::Tokenizer::new(&document,
|
||||||
Tokenizer::Html(
|
url.clone(),
|
||||||
self::html::Tokenizer::new(&document, url.clone(), Some(fragment_context))),
|
Some(fragment_context))),
|
||||||
LastChunkState::Received);
|
LastChunkState::Received);
|
||||||
parser.parse_chunk(String::from(input));
|
parser.parse_chunk(String::from(input));
|
||||||
|
|
||||||
// Step 14.
|
// Step 14.
|
||||||
let root_element = document.GetDocumentElement().expect("no document element");
|
let root_element = document.GetDocumentElement().expect("no document element");
|
||||||
FragmentParsingResult { inner: root_element.upcast::<Node>().children() }
|
FragmentParsingResult {
|
||||||
|
inner: root_element.upcast::<Node>().children(),
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn parse_xml_document(
|
pub fn parse_xml_document(document: &Document, input: DOMString, url: ServoUrl, owner: Option<PipelineId>) {
|
||||||
document: &Document,
|
let parser = ServoParser::new(document,
|
||||||
input: DOMString,
|
owner,
|
||||||
url: ServoUrl,
|
Tokenizer::Xml(self::xml::Tokenizer::new(document, url)),
|
||||||
owner: Option<PipelineId>) {
|
LastChunkState::NotReceived);
|
||||||
let parser = ServoParser::new(
|
|
||||||
document,
|
|
||||||
owner,
|
|
||||||
Tokenizer::Xml(self::xml::Tokenizer::new(document, url)),
|
|
||||||
LastChunkState::NotReceived);
|
|
||||||
parser.parse_chunk(String::from(input));
|
parser.parse_chunk(String::from(input));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -179,7 +175,8 @@ impl ServoParser {
|
||||||
assert!(self.suspended.get());
|
assert!(self.suspended.get());
|
||||||
self.suspended.set(false);
|
self.suspended.set(false);
|
||||||
|
|
||||||
mem::swap(&mut *self.script_input.borrow_mut(), &mut *self.network_input.borrow_mut());
|
mem::swap(&mut *self.script_input.borrow_mut(),
|
||||||
|
&mut *self.network_input.borrow_mut());
|
||||||
while let Some(chunk) = self.script_input.borrow_mut().pop_front() {
|
while let Some(chunk) = self.script_input.borrow_mut().pop_front() {
|
||||||
self.network_input.borrow_mut().push_back(chunk);
|
self.network_input.borrow_mut().push_back(chunk);
|
||||||
}
|
}
|
||||||
|
@ -236,12 +233,11 @@ impl ServoParser {
|
||||||
}
|
}
|
||||||
|
|
||||||
#[allow(unrooted_must_root)]
|
#[allow(unrooted_must_root)]
|
||||||
fn new_inherited(
|
fn new_inherited(document: &Document,
|
||||||
document: &Document,
|
pipeline: Option<PipelineId>,
|
||||||
pipeline: Option<PipelineId>,
|
tokenizer: Tokenizer,
|
||||||
tokenizer: Tokenizer,
|
last_chunk_state: LastChunkState)
|
||||||
last_chunk_state: LastChunkState)
|
-> Self {
|
||||||
-> Self {
|
|
||||||
ServoParser {
|
ServoParser {
|
||||||
reflector: Reflector::new(),
|
reflector: Reflector::new(),
|
||||||
document: JS::from_ref(document),
|
document: JS::from_ref(document),
|
||||||
|
@ -256,16 +252,14 @@ impl ServoParser {
|
||||||
}
|
}
|
||||||
|
|
||||||
#[allow(unrooted_must_root)]
|
#[allow(unrooted_must_root)]
|
||||||
fn new(
|
fn new(document: &Document,
|
||||||
document: &Document,
|
pipeline: Option<PipelineId>,
|
||||||
pipeline: Option<PipelineId>,
|
tokenizer: Tokenizer,
|
||||||
tokenizer: Tokenizer,
|
last_chunk_state: LastChunkState)
|
||||||
last_chunk_state: LastChunkState)
|
-> Root<Self> {
|
||||||
-> Root<Self> {
|
reflect_dom_object(box ServoParser::new_inherited(document, pipeline, tokenizer, last_chunk_state),
|
||||||
reflect_dom_object(
|
document.window(),
|
||||||
box ServoParser::new_inherited(document, pipeline, tokenizer, last_chunk_state),
|
ServoParserBinding::Wrap)
|
||||||
document.window(),
|
|
||||||
ServoParserBinding::Wrap)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
fn push_input_chunk(&self, chunk: String) {
|
fn push_input_chunk(&self, chunk: String) {
|
||||||
|
@ -313,7 +307,7 @@ impl ServoParser {
|
||||||
}
|
}
|
||||||
|
|
||||||
fn tokenize<F>(&self, mut feed: F)
|
fn tokenize<F>(&self, mut feed: F)
|
||||||
where F: FnMut(&mut Tokenizer) -> Result<(), Root<HTMLScriptElement>>
|
where F: FnMut(&mut Tokenizer) -> Result<(), Root<HTMLScriptElement>>,
|
||||||
{
|
{
|
||||||
loop {
|
loop {
|
||||||
assert!(!self.suspended.get());
|
assert!(!self.suspended.get());
|
||||||
|
@ -356,7 +350,9 @@ impl ServoParser {
|
||||||
self.document.disarm_reflow_timeout();
|
self.document.disarm_reflow_timeout();
|
||||||
self.document.upcast::<Node>().dirty(NodeDamage::OtherNodeDamage);
|
self.document.upcast::<Node>().dirty(NodeDamage::OtherNodeDamage);
|
||||||
let window = self.document.window();
|
let window = self.document.window();
|
||||||
window.reflow(ReflowGoal::ForDisplay, ReflowQueryType::NoQuery, ReflowReason::FirstLoad);
|
window.reflow(ReflowGoal::ForDisplay,
|
||||||
|
ReflowQueryType::NoQuery,
|
||||||
|
ReflowReason::FirstLoad);
|
||||||
}
|
}
|
||||||
|
|
||||||
// Steps 3-12 are in another castle, namely finish_load.
|
// Steps 3-12 are in another castle, namely finish_load.
|
||||||
|
@ -455,14 +451,13 @@ impl FetchResponseListener for ParserContext {
|
||||||
|
|
||||||
fn process_request_eof(&mut self) {}
|
fn process_request_eof(&mut self) {}
|
||||||
|
|
||||||
fn process_response(&mut self,
|
fn process_response(&mut self, meta_result: Result<FetchMetadata, NetworkError>) {
|
||||||
meta_result: Result<FetchMetadata, NetworkError>) {
|
|
||||||
let mut ssl_error = None;
|
let mut ssl_error = None;
|
||||||
let metadata = match meta_result {
|
let metadata = match meta_result {
|
||||||
Ok(meta) => {
|
Ok(meta) => {
|
||||||
Some(match meta {
|
Some(match meta {
|
||||||
FetchMetadata::Unfiltered(m) => m,
|
FetchMetadata::Unfiltered(m) => m,
|
||||||
FetchMetadata::Filtered { unsafe_, .. } => unsafe_
|
FetchMetadata::Filtered { unsafe_, .. } => unsafe_,
|
||||||
})
|
})
|
||||||
},
|
},
|
||||||
Err(NetworkError::SslValidation(url, reason)) => {
|
Err(NetworkError::SslValidation(url, reason)) => {
|
||||||
|
@ -474,10 +469,8 @@ impl FetchResponseListener for ParserContext {
|
||||||
},
|
},
|
||||||
Err(_) => None,
|
Err(_) => None,
|
||||||
};
|
};
|
||||||
let content_type =
|
let content_type = metadata.clone().and_then(|meta| meta.content_type).map(Serde::into_inner);
|
||||||
metadata.clone().and_then(|meta| meta.content_type).map(Serde::into_inner);
|
let parser = match ScriptThread::page_headers_available(&self.id, metadata) {
|
||||||
let parser = match ScriptThread::page_headers_available(&self.id,
|
|
||||||
metadata) {
|
|
||||||
Some(parser) => parser,
|
Some(parser) => parser,
|
||||||
None => return,
|
None => return,
|
||||||
};
|
};
|
||||||
|
@ -505,7 +498,8 @@ impl FetchResponseListener for ParserContext {
|
||||||
parser.parse_sync();
|
parser.parse_sync();
|
||||||
parser.tokenizer.borrow_mut().set_plaintext_state();
|
parser.tokenizer.borrow_mut().set_plaintext_state();
|
||||||
},
|
},
|
||||||
Some(ContentType(Mime(TopLevel::Text, SubLevel::Html, _))) => { // Handle text/html
|
Some(ContentType(Mime(TopLevel::Text, SubLevel::Html, _))) => {
|
||||||
|
// Handle text/html
|
||||||
if let Some(reason) = ssl_error {
|
if let Some(reason) = ssl_error {
|
||||||
self.is_synthesized_document = true;
|
self.is_synthesized_document = true;
|
||||||
let page_bytes = read_resource_file("badcert.html").unwrap();
|
let page_bytes = read_resource_file("badcert.html").unwrap();
|
||||||
|
@ -524,7 +518,8 @@ impl FetchResponseListener for ParserContext {
|
||||||
|
|
||||||
// Show warning page for unknown mime types.
|
// Show warning page for unknown mime types.
|
||||||
let page = format!("<html><body><p>Unknown content type ({}/{}).</p></body></html>",
|
let page = format!("<html><body><p>Unknown content type ({}/{}).</p></body></html>",
|
||||||
toplevel.as_str(), sublevel.as_str());
|
toplevel.as_str(),
|
||||||
|
sublevel.as_str());
|
||||||
self.is_synthesized_document = true;
|
self.is_synthesized_document = true;
|
||||||
parser.push_input_chunk(page);
|
parser.push_input_chunk(page);
|
||||||
parser.parse_sync();
|
parser.parse_sync();
|
||||||
|
@ -532,7 +527,7 @@ impl FetchResponseListener for ParserContext {
|
||||||
None => {
|
None => {
|
||||||
// No content-type header.
|
// No content-type header.
|
||||||
// Merge with #4212 when fixed.
|
// Merge with #4212 when fixed.
|
||||||
}
|
},
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -595,6 +590,6 @@ fn insert(parent: &Node, reference_child: Option<&Node>, child: NodeOrText<JS<No
|
||||||
let text = Text::new(String::from(t).into(), &parent.owner_doc());
|
let text = Text::new(String::from(t).into(), &parent.owner_doc());
|
||||||
parent.InsertBefore(text.upcast(), reference_child).unwrap();
|
parent.InsertBefore(text.upcast(), reference_child).unwrap();
|
||||||
}
|
}
|
||||||
}
|
},
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue