mirror of
https://github.com/servo/servo.git
synced 2025-06-06 16:45:39 +00:00
Add a NetworkEventActor to devtools/actors/ Authors: Ashritha Mohan Ram <amohanr@ncsu.edu> Himaja Valavala <hsvalava@ncsu.edu> Anand Chandrasekar <achandr9@ncsu.edu> Yiyang Wang <ywang95@ncsu.edu>
404 lines
16 KiB
Rust
404 lines
16 KiB
Rust
/* This Source Code Form is subject to the terms of the Mozilla Public
|
|
* License, v. 2.0. If a copy of the MPL was not distributed with this
|
|
* file, You can obtain one at http://mozilla.org/MPL/2.0/. */
|
|
|
|
use net_traits::{ControlMsg, CookieSource, LoadData, Metadata, LoadConsumer};
|
|
use net_traits::ProgressMsg::{Payload, Done};
|
|
use devtools_traits::{DevtoolsControlMsg};
|
|
use mime_classifier::MIMEClassifier;
|
|
use resource_task::{start_sending_opt, start_sending_sniffed_opt};
|
|
|
|
use log;
|
|
use std::collections::HashSet;
|
|
use file_loader;
|
|
use flate2::read::{DeflateDecoder, GzDecoder};
|
|
use hyper::client::Request;
|
|
use hyper::header::{AcceptEncoding, Accept, ContentLength, ContentType, Host, Location, qitem, Quality, QualityItem};
|
|
use hyper::HttpError;
|
|
use hyper::method::Method;
|
|
use hyper::mime::{Mime, TopLevel, SubLevel};
|
|
use hyper::net::HttpConnector;
|
|
use hyper::status::{StatusCode, StatusClass};
|
|
use std::error::Error;
|
|
use openssl::ssl::{SslContext, SSL_VERIFY_PEER};
|
|
use std::io::{self, Read, Write};
|
|
use std::sync::Arc;
|
|
use std::sync::mpsc::{Sender, channel};
|
|
use util::task::spawn_named;
|
|
use util::resource_files::resources_dir_path;
|
|
use util::opts;
|
|
use url::{Url, UrlParser};
|
|
|
|
use std::borrow::ToOwned;
|
|
use std::boxed::FnBox;
|
|
|
|
pub fn factory(cookies_chan: Sender<ControlMsg>, devtools_chan: Option<Sender<DevtoolsControlMsg>>)
|
|
-> Box<FnBox(LoadData, LoadConsumer, Arc<MIMEClassifier>) + Send> {
|
|
box move |load_data, senders, classifier| {
|
|
spawn_named("http_loader".to_owned(), move || load(load_data, senders, classifier, cookies_chan, devtools_chan))
|
|
}
|
|
}
|
|
|
|
fn send_error(url: Url, err: String, start_chan: LoadConsumer) {
|
|
let mut metadata: Metadata = Metadata::default(url);
|
|
metadata.status = None;
|
|
|
|
match start_sending_opt(start_chan, metadata) {
|
|
Ok(p) => p.send(Done(Err(err))).unwrap(),
|
|
_ => {}
|
|
};
|
|
}
|
|
|
|
enum ReadResult {
|
|
Payload(Vec<u8>),
|
|
EOF,
|
|
}
|
|
|
|
fn read_block<R: Read>(reader: &mut R) -> Result<ReadResult, ()> {
|
|
let mut buf = vec![0; 1024];
|
|
|
|
match reader.read(&mut buf) {
|
|
Ok(len) if len > 0 => {
|
|
unsafe { buf.set_len(len); }
|
|
Ok(ReadResult::Payload(buf))
|
|
}
|
|
Ok(_) => Ok(ReadResult::EOF),
|
|
Err(_) => Err(()),
|
|
}
|
|
}
|
|
|
|
fn load(mut load_data: LoadData, start_chan: LoadConsumer, classifier: Arc<MIMEClassifier>,
|
|
cookies_chan: Sender<ControlMsg>, devtools_chan: Option<Sender<DevtoolsControlMsg>>) {
|
|
// FIXME: At the time of writing this FIXME, servo didn't have any central
|
|
// location for configuration. If you're reading this and such a
|
|
// repository DOES exist, please update this constant to use it.
|
|
let max_redirects = 50;
|
|
let mut iters = 0;
|
|
let mut url = load_data.url.clone();
|
|
let mut redirected_to = HashSet::new();
|
|
|
|
// If the URL is a view-source scheme then the scheme data contains the
|
|
// real URL that should be used for which the source is to be viewed.
|
|
// Change our existing URL to that and keep note that we are viewing
|
|
// the source rather than rendering the contents of the URL.
|
|
let viewing_source = url.scheme == "view-source";
|
|
if viewing_source {
|
|
let inner_url = load_data.url.non_relative_scheme_data().unwrap();
|
|
url = Url::parse(inner_url).unwrap();
|
|
match &*url.scheme {
|
|
"http" | "https" => {}
|
|
_ => {
|
|
let s = format!("The {} scheme with view-source is not supported", url.scheme);
|
|
send_error(url, s, start_chan);
|
|
return;
|
|
}
|
|
};
|
|
}
|
|
|
|
// Loop to handle redirects.
|
|
loop {
|
|
iters = iters + 1;
|
|
|
|
if iters > max_redirects {
|
|
send_error(url, "too many redirects".to_string(), start_chan);
|
|
return;
|
|
}
|
|
|
|
match &*url.scheme {
|
|
"http" | "https" => {}
|
|
_ => {
|
|
let s = format!("{} request, but we don't support that scheme", url.scheme);
|
|
send_error(url, s, start_chan);
|
|
return;
|
|
}
|
|
}
|
|
|
|
info!("requesting {}", url.serialize());
|
|
|
|
fn verifier(ssl: &mut SslContext) {
|
|
ssl.set_verify(SSL_VERIFY_PEER, None);
|
|
let mut certs = resources_dir_path();
|
|
certs.push("certs");
|
|
ssl.set_CA_file(&certs).unwrap();
|
|
};
|
|
|
|
let ssl_err_string = "Some(OpenSslErrors([UnknownError { library: \"SSL routines\", \
|
|
function: \"SSL3_GET_SERVER_CERTIFICATE\", \
|
|
reason: \"certificate verify failed\" }]))";
|
|
|
|
let mut connector = if opts::get().nossl {
|
|
HttpConnector(None)
|
|
} else {
|
|
HttpConnector(Some(box verifier as Box<FnMut(&mut SslContext) + Send>))
|
|
};
|
|
|
|
let mut req = match Request::with_connector(load_data.method.clone(), url.clone(), &mut connector) {
|
|
Ok(req) => req,
|
|
Err(HttpError::HttpIoError(ref io_error)) if (
|
|
io_error.kind() == io::ErrorKind::Other &&
|
|
io_error.description() == "Error in OpenSSL" &&
|
|
// FIXME: This incredibly hacky. Make it more robust, and at least test it.
|
|
format!("{:?}", io_error.cause()) == ssl_err_string
|
|
) => {
|
|
let mut image = resources_dir_path();
|
|
image.push("badcert.html");
|
|
let load_data = LoadData::new(Url::from_file_path(&*image).unwrap());
|
|
file_loader::factory(load_data, start_chan, classifier);
|
|
return;
|
|
},
|
|
Err(e) => {
|
|
println!("{:?}", e);
|
|
send_error(url, e.description().to_string(), start_chan);
|
|
return;
|
|
}
|
|
};
|
|
|
|
// Preserve the `host` header set automatically by Request.
|
|
let host = req.headers().get::<Host>().unwrap().clone();
|
|
|
|
// Avoid automatically preserving request headers when redirects occur.
|
|
// See https://bugzilla.mozilla.org/show_bug.cgi?id=401564 and
|
|
// https://bugzilla.mozilla.org/show_bug.cgi?id=216828 .
|
|
// Only preserve ones which have been explicitly marked as such.
|
|
if iters == 1 {
|
|
let mut combined_headers = load_data.headers.clone();
|
|
combined_headers.extend(load_data.preserved_headers.iter());
|
|
*req.headers_mut() = combined_headers;
|
|
} else {
|
|
*req.headers_mut() = load_data.preserved_headers.clone();
|
|
}
|
|
|
|
req.headers_mut().set(host);
|
|
|
|
if !req.headers().has::<Accept>() {
|
|
let accept = Accept(vec![
|
|
qitem(Mime(TopLevel::Text, SubLevel::Html, vec![])),
|
|
qitem(Mime(TopLevel::Application, SubLevel::Ext("xhtml+xml".to_string()), vec![])),
|
|
QualityItem::new(Mime(TopLevel::Application, SubLevel::Xml, vec![]), Quality(900u16)),
|
|
QualityItem::new(Mime(TopLevel::Star, SubLevel::Star, vec![]), Quality(800u16)),
|
|
]);
|
|
req.headers_mut().set(accept);
|
|
}
|
|
|
|
let (tx, rx) = channel();
|
|
cookies_chan.send(ControlMsg::GetCookiesForUrl(url.clone(), tx, CookieSource::HTTP)).unwrap();
|
|
if let Some(cookie_list) = rx.recv().unwrap() {
|
|
let mut v = Vec::new();
|
|
v.push(cookie_list.into_bytes());
|
|
req.headers_mut().set_raw("Cookie".to_owned(), v);
|
|
}
|
|
|
|
if !req.headers().has::<AcceptEncoding>() {
|
|
req.headers_mut().set_raw("Accept-Encoding".to_owned(), vec![b"gzip, deflate".to_vec()]);
|
|
}
|
|
if log_enabled!(log::INFO) {
|
|
info!("{}", load_data.method);
|
|
for header in req.headers().iter() {
|
|
info!(" - {}", header);
|
|
}
|
|
info!("{:?}", load_data.data);
|
|
}
|
|
|
|
/*
|
|
match devtools_chan {
|
|
Some(chan) => chan.send(DevtoolsControlMsg::HttpRequest(load_data.url.clone(), load_data.method.clone(), load_data.headers.clone(), load_data.data.clone())).unwrap(),
|
|
None => {}
|
|
}
|
|
*/
|
|
|
|
println!("load");
|
|
devtools_chan.as_ref().map(|chan| chan.send(DevtoolsControlMsg::HttpRequest(load_data.url.clone(), load_data.method.clone(), load_data.headers.clone(), load_data.data.clone())).unwrap());
|
|
|
|
|
|
// Avoid automatically sending request body if a redirect has occurred.
|
|
let writer = match load_data.data {
|
|
Some(ref data) if iters == 1 => {
|
|
req.headers_mut().set(ContentLength(data.len() as u64));
|
|
let mut writer = match req.start() {
|
|
Ok(w) => w,
|
|
Err(e) => {
|
|
send_error(url, e.description().to_string(), start_chan);
|
|
return;
|
|
}
|
|
};
|
|
match writer.write_all(&*data) {
|
|
Err(e) => {
|
|
send_error(url, e.description().to_string(), start_chan);
|
|
return;
|
|
}
|
|
_ => {}
|
|
};
|
|
writer
|
|
},
|
|
_ => {
|
|
match load_data.method {
|
|
Method::Get | Method::Head => (),
|
|
_ => req.headers_mut().set(ContentLength(0))
|
|
}
|
|
match req.start() {
|
|
Ok(w) => w,
|
|
Err(e) => {
|
|
send_error(url, e.description().to_string(), start_chan);
|
|
return;
|
|
}
|
|
}
|
|
}
|
|
};
|
|
let mut response = match writer.send() {
|
|
Ok(r) => r,
|
|
Err(e) => {
|
|
send_error(url, e.description().to_string(), start_chan);
|
|
return;
|
|
}
|
|
};
|
|
|
|
// Dump headers, but only do the iteration if info!() is enabled.
|
|
info!("got HTTP response {}, headers:", response.status);
|
|
if log_enabled!(log::INFO) {
|
|
for header in response.headers.iter() {
|
|
info!(" - {}", header);
|
|
}
|
|
}
|
|
|
|
if let Some(cookies) = response.headers.get_raw("set-cookie") {
|
|
for cookie in cookies.iter() {
|
|
if let Ok(cookies) = String::from_utf8(cookie.clone()) {
|
|
cookies_chan.send(ControlMsg::SetCookiesForUrl(url.clone(),
|
|
cookies,
|
|
CookieSource::HTTP)).unwrap();
|
|
}
|
|
}
|
|
}
|
|
|
|
if response.status.class() == StatusClass::Redirection {
|
|
match response.headers.get::<Location>() {
|
|
Some(&Location(ref new_url)) => {
|
|
// CORS (https://fetch.spec.whatwg.org/#http-fetch, status section, point 9, 10)
|
|
match load_data.cors {
|
|
Some(ref c) => {
|
|
if c.preflight {
|
|
// The preflight lied
|
|
send_error(url, "Preflight fetch inconsistent with main fetch".to_string(), start_chan);
|
|
return;
|
|
} else {
|
|
// XXXManishearth There are some CORS-related steps here,
|
|
// but they don't seem necessary until credentials are implemented
|
|
}
|
|
}
|
|
_ => {}
|
|
}
|
|
let new_url = match UrlParser::new().base_url(&url).parse(&new_url) {
|
|
Ok(u) => u,
|
|
Err(e) => {
|
|
send_error(url, e.to_string(), start_chan);
|
|
return;
|
|
}
|
|
};
|
|
info!("redirecting to {}", new_url);
|
|
url = new_url;
|
|
|
|
// According to https://tools.ietf.org/html/rfc7231#section-6.4.2,
|
|
// historically UAs have rewritten POST->GET on 301 and 302 responses.
|
|
if load_data.method == Method::Post &&
|
|
(response.status == StatusCode::MovedPermanently ||
|
|
response.status == StatusCode::Found) {
|
|
load_data.method = Method::Get;
|
|
}
|
|
|
|
if redirected_to.contains(&url) {
|
|
send_error(url, "redirect loop".to_string(), start_chan);
|
|
return;
|
|
}
|
|
|
|
redirected_to.insert(url.clone());
|
|
continue;
|
|
}
|
|
None => ()
|
|
}
|
|
}
|
|
|
|
let mut adjusted_headers = response.headers.clone();
|
|
if viewing_source {
|
|
adjusted_headers.set(ContentType(Mime(TopLevel::Text, SubLevel::Plain, vec![])));
|
|
}
|
|
let mut metadata: Metadata = Metadata::default(url);
|
|
metadata.set_content_type(match adjusted_headers.get() {
|
|
Some(&ContentType(ref mime)) => Some(mime),
|
|
None => None
|
|
});
|
|
metadata.headers = Some(adjusted_headers);
|
|
metadata.status = Some(response.status_raw().clone());
|
|
|
|
let mut encoding_str: Option<String> = None;
|
|
//FIXME: Implement Content-Encoding Header https://github.com/hyperium/hyper/issues/391
|
|
if let Some(encodings) = response.headers.get_raw("content-encoding") {
|
|
for encoding in encodings.iter() {
|
|
if let Ok(encodings) = String::from_utf8(encoding.clone()) {
|
|
if encodings == "gzip" || encodings == "deflate" {
|
|
encoding_str = Some(encodings);
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
match encoding_str {
|
|
Some(encoding) => {
|
|
if encoding == "gzip" {
|
|
let result = GzDecoder::new(response);
|
|
match result {
|
|
Ok(mut response_decoding) => {
|
|
send_data(&mut response_decoding, start_chan, metadata, classifier);
|
|
}
|
|
Err(err) => {
|
|
send_error(metadata.final_url, err.to_string(), start_chan);
|
|
return;
|
|
}
|
|
}
|
|
} else if encoding == "deflate" {
|
|
let mut response_decoding = DeflateDecoder::new(response);
|
|
send_data(&mut response_decoding, start_chan, metadata, classifier);
|
|
}
|
|
},
|
|
None => {
|
|
send_data(&mut response, start_chan, metadata, classifier);
|
|
}
|
|
}
|
|
|
|
// We didn't get redirected.
|
|
break;
|
|
}
|
|
}
|
|
|
|
fn send_data<R: Read>(reader: &mut R,
|
|
start_chan: LoadConsumer,
|
|
metadata: Metadata,
|
|
classifier: Arc<MIMEClassifier>) {
|
|
let (progress_chan, mut chunk) = {
|
|
let buf = match read_block(reader) {
|
|
Ok(ReadResult::Payload(buf)) => buf,
|
|
_ => vec!(),
|
|
};
|
|
let p = match start_sending_sniffed_opt(start_chan, metadata, classifier, &buf) {
|
|
Ok(p) => p,
|
|
_ => return
|
|
};
|
|
(p, buf)
|
|
};
|
|
|
|
loop {
|
|
if progress_chan.send(Payload(chunk)).is_err() {
|
|
// The send errors when the receiver is out of scope,
|
|
// which will happen if the fetch has timed out (or has been aborted)
|
|
// so we don't need to continue with the loading of the file here.
|
|
return;
|
|
}
|
|
|
|
chunk = match read_block(reader) {
|
|
Ok(ReadResult::Payload(buf)) => buf,
|
|
Ok(ReadResult::EOF) | Err(_) => break,
|
|
};
|
|
}
|
|
|
|
let _ = progress_chan.send(Done(Ok(())));
|
|
}
|