Finish asynchronous blob url fetching

This commit is contained in:
Fernando Jiménez Moreno 2018-11-07 15:00:49 +01:00
parent 67722d1943
commit 8538634210
3 changed files with 147 additions and 136 deletions

View file

@ -2,20 +2,13 @@
* License, v. 2.0. If a copy of the MPL was not distributed with this * License, v. 2.0. If a copy of the MPL was not distributed with this
* file, You can obtain one at https://mozilla.org/MPL/2.0/. */ * file, You can obtain one at https://mozilla.org/MPL/2.0/. */
use crate::fetch::methods::{Data, DoneChannel};
use crate::filemanager_thread::FileManager; use crate::filemanager_thread::FileManager;
use fetch::methods::DoneChannel;
use headers_core::HeaderMapExt;
use headers_ext::{ContentLength, ContentType};
use http::header::{self, HeaderValue};
use http::HeaderMap;
use ipc_channel::ipc;
use mime::{self, Mime};
use net_traits::blob_url_store::parse_blob_url; use net_traits::blob_url_store::parse_blob_url;
use net_traits::filemanager_thread::ReadFileProgress;
use net_traits::response::{Response, ResponseBody}; use net_traits::response::{Response, ResponseBody};
use net_traits::{http_percent_encode, NetworkError, ResourceFetchTiming}; use net_traits::{NetworkError, ResourceFetchTiming};
use servo_url::ServoUrl;
use servo_channel::channel; use servo_channel::channel;
use servo_url::ServoUrl;
// TODO: Check on GET // TODO: Check on GET
// https://w3c.github.io/FileAPI/#requestResponseModel // https://w3c.github.io/FileAPI/#requestResponseModel
@ -24,13 +17,13 @@ use servo_channel::channel;
pub fn load_blob_async( pub fn load_blob_async(
url: ServoUrl, url: ServoUrl,
filemanager: FileManager, filemanager: FileManager,
done_chan: &mut DoneChannel done_chan: &mut DoneChannel,
)-> Response { ) -> Response {
let (id, origin) = match parse_blob_url(&url) { let (id, origin) = match parse_blob_url(&url) {
Ok((id, origin)) => (id, origin), Ok((id, origin)) => (id, origin),
Err(()) => { Err(()) => {
return Response::network_error(NetworkError::Internal("Invalid blob url".into())); return Response::network_error(NetworkError::Internal("Invalid blob url".into()));
} },
}; };
let mut response = Response::new(url, ResourceFetchTiming::new(request.timing_type())); let mut response = Response::new(url, ResourceFetchTiming::new(request.timing_type()));
@ -38,7 +31,11 @@ pub fn load_blob_async(
*done_chan = Some((sender.clone(), receiver)); *done_chan = Some((sender.clone(), receiver));
*response.body.lock().unwrap() = ResponseBody::Receiving(vec![]); *response.body.lock().unwrap() = ResponseBody::Receiving(vec![]);
let check_url_validity = true; let check_url_validity = true;
filemanager.fetch_file(sender, id, check_url_validity, origin, &mut response); if let Err(err) = filemanager.fetch_file(&sender, id, check_url_validity, origin, &mut response)
{
let _ = sender.send(Data::Done);
return Response::network_error(NetworkError::Internal(err));
};
response response
} }

View file

@ -657,7 +657,7 @@ fn scheme_fetch(
)); ));
} }
load_blob_async(url.clone(), context.filemanager.clone(), done_chan); load_blob_async(url.clone(), context.filemanager.clone(), done_chan)
}, },
"ftp" => { "ftp" => {

View file

@ -2,22 +2,21 @@
* License, v. 2.0. If a copy of the MPL was not distributed with this * License, v. 2.0. If a copy of the MPL was not distributed with this
* file, You can obtain one at https://mozilla.org/MPL/2.0/. */ * file, You can obtain one at https://mozilla.org/MPL/2.0/. */
use crate::fetch::methods::Data;
use embedder_traits::{EmbedderMsg, EmbedderProxy, FilterPattern}; use embedder_traits::{EmbedderMsg, EmbedderProxy, FilterPattern};
use ipc_channel::ipc::{self, IpcSender}; use headers_ext::{ContentLength, ContentType, HeaderMap, HeaderMapExt};
use fetch::methods::Data;
use headers_core::HeaderMapExt;
use headers_ext::{ContentLength, ContentType};
use http::HeaderMap;
use http::header::{self, HeaderValue}; use http::header::{self, HeaderValue};
use ipc_channel::ipc::{self, IpcSender};
use mime::{self, Mime}; use mime::{self, Mime};
use mime_guess::guess_mime_type_opt; use mime_guess::guess_mime_type_opt;
use net_traits::{http_percent_encode, NetworkError};
use net_traits::blob_url_store::{BlobBuf, BlobURLStoreError}; use net_traits::blob_url_store::{BlobBuf, BlobURLStoreError};
use net_traits::filemanager_thread::{FileManagerResult, FileManagerThreadMsg, FileOrigin}; use net_traits::filemanager_thread::{FileManagerResult, FileManagerThreadMsg, FileOrigin};
use net_traits::filemanager_thread::{ use net_traits::filemanager_thread::{
FileManagerThreadError, ReadFileProgress, RelativePos, SelectedFile, FileManagerThreadError, ReadFileProgress, RelativePos, SelectedFile,
}; };
use net_traits::http_percent_encode;
use net_traits::response::{Response, ResponseBody}; use net_traits::response::{Response, ResponseBody};
use servo_arc::Arc as ServoArc;
use servo_channel; use servo_channel;
use servo_config::prefs::PREFS; use servo_config::prefs::PREFS;
use std::collections::HashMap; use std::collections::HashMap;
@ -26,7 +25,7 @@ use std::io::{Read, Seek, SeekFrom};
use std::ops::Index; use std::ops::Index;
use std::path::{Path, PathBuf}; use std::path::{Path, PathBuf};
use std::sync::atomic::{self, AtomicBool, AtomicUsize, Ordering}; use std::sync::atomic::{self, AtomicBool, AtomicUsize, Ordering};
use std::sync::{Arc, RwLock}; use std::sync::{Arc, Mutex, RwLock};
use std::thread; use std::thread;
use url::Url; use url::Url;
use uuid::Uuid; use uuid::Uuid;
@ -100,17 +99,27 @@ impl FileManager {
.expect("Thread spawning failed"); .expect("Thread spawning failed");
} }
pub fn fetch_file(&self, // Read a file for the Fetch implementation.
sender: servo_channel::Sender<Data>, // It gets the required headers synchronously and reads the actual content
// in a separate thread.
pub fn fetch_file(
&self,
sender: &servo_channel::Sender<Data>,
id: Uuid, id: Uuid,
check_url_validity: bool, check_url_validity: bool,
origin: FileOrigin, origin: FileOrigin,
response: &mut Response) { response: &mut Response,
let store = self.store.clone(); ) -> Result<(), String> {
let mut r2 = response.clone(); self.store
thread::Builder::new().name("read file".to_owned()).spawn(move || { .fetch_blob_buf(
store.try_fetch_file(&sender, id, check_url_validity, origin, &mut r2) sender,
}).expect("Thread spawning failed"); &id,
&origin,
RelativePos::full_range(),
check_url_validity,
response,
)
.map_err(|e| format!("{:?}", e))
} }
pub fn promote_memory( pub fn promote_memory(
@ -511,52 +520,36 @@ impl FileManagerStore {
) )
} }
fn fetch_blob_buf(&self, sender: &servo_channel::Sender<Data>, fn fetch_blob_buf(
id: &Uuid, origin_in: &FileOrigin, rel_pos: RelativePos, &self,
check_url_validity: bool, response: &mut Response) -> Result<(), BlobURLStoreError> { sender: &servo_channel::Sender<Data>,
let mut bytes = vec![]; id: &Uuid,
origin_in: &FileOrigin,
rel_pos: RelativePos,
check_url_validity: bool,
response: &mut Response,
) -> Result<(), BlobURLStoreError> {
let file_impl = self.get_impl(id, origin_in, check_url_validity)?; let file_impl = self.get_impl(id, origin_in, check_url_validity)?;
match file_impl { match file_impl {
FileImpl::Memory(buf) => { FileImpl::Memory(buf) => {
let range = rel_pos.to_abs_range(buf.size as usize); let range = rel_pos.to_abs_range(buf.size as usize);
let blob_buf = BlobBuf { let len = range.len() as u64;
filename: None,
type_string: buf.type_string,
size: range.len() as u64,
bytes: buf.bytes.index(range).to_vec(),
};
let content_type: Mime = blob_buf.type_string.parse().unwrap_or(mime::TEXT_PLAIN); set_headers(
let charset = content_type.get_param(mime::CHARSET); &mut response.headers,
let mut headers = HeaderMap::new(); len,
buf.type_string.parse().unwrap_or(mime::TEXT_PLAIN),
if let Some(name) = blob_buf.filename { /* filename */ None,
let charset = charset.map(|c| c.as_ref().into()).unwrap_or("us-ascii".to_owned());
// TODO(eijebong): Replace this once the typed header is there
headers.insert(
header::CONTENT_DISPOSITION,
HeaderValue::from_bytes(
format!("inline; {}",
if charset.to_lowercase() == "utf-8" {
format!("filename=\"{}\"", String::from_utf8(name.as_bytes().into()).unwrap())
} else {
format!("filename*=\"{}\"''{}", charset, http_percent_encode(name.as_bytes()))
}
).as_bytes()
).unwrap()
); );
}
headers.typed_insert(ContentLength(blob_buf.size as u64)); let mut bytes = vec![];
headers.typed_insert(ContentType::from(content_type.clone())); bytes.extend_from_slice(buf.bytes.index(range));
bytes.extend_from_slice(&blob_buf.bytes); let _ = sender.send(Data::Payload(bytes));
response.headers = headers;
*response.body.lock().unwrap() = ResponseBody::Done(bytes);
let _ = sender.send(Data::Done); let _ = sender.send(Data::Done);
Ok(()) Ok(())
} },
FileImpl::MetaDataOnly(metadata) => { FileImpl::MetaDataOnly(metadata) => {
/* XXX: Snapshot state check (optional) https://w3c.github.io/FileAPI/#snapshot-state. /* XXX: Snapshot state check (optional) https://w3c.github.io/FileAPI/#snapshot-state.
Concretely, here we create another file, and this file might not Concretely, here we create another file, and this file might not
@ -564,45 +557,53 @@ impl FileManagerStore {
create_entry is called. create_entry is called.
*/ */
let opt_filename = metadata.path.file_name() let mut file = File::open(&metadata.path)
.map_err(|e| BlobURLStoreError::External(e.to_string()))?;
let range = rel_pos.to_abs_range(metadata.size as usize);
let range_start = range.start as u64;
let seeked_start = file
.seek(SeekFrom::Start(range_start))
.map_err(|e| BlobURLStoreError::External(e.to_string()))?;
if seeked_start != range_start {
return Err(BlobURLStoreError::InvalidEntry);
}
let filename = metadata
.path
.file_name()
.and_then(|osstr| osstr.to_str()) .and_then(|osstr| osstr.to_str())
.map(|s| s.to_string()); .map(|s| s.to_string());
let mime = guess_mime_type_opt(metadata.path.clone()); set_headers(
let range = rel_pos.to_abs_range(metadata.size as usize); &mut response.headers,
metadata.size,
guess_mime_type_opt(metadata.path).unwrap_or(mime::TEXT_PLAIN),
filename,
);
let mut file = File::open(&metadata.path) let body = response.body.clone();
.map_err(|e| BlobURLStoreError::External(e.to_string()))?; let sender = sender.clone();
let seeked_start = file.seek(SeekFrom::Start(range.start as u64)) thread::Builder::new()
.map_err(|e| BlobURLStoreError::External(e.to_string()))?; .name("fetch file".to_owned())
.spawn(move || chunked_fetch(sender, &mut file, body))
if seeked_start == (range.start as u64) { .expect("Thread spawn failed");
let type_string = match mime {
Some(x) => format!("{}", x),
None => "".to_string(),
};
chunked_fetch(sender, &mut file, range.len(), opt_filename,
type_string, response, &mut bytes);
Ok(()) Ok(())
} else { },
Err(BlobURLStoreError::InvalidEntry)
}
}
FileImpl::Sliced(parent_id, inner_rel_pos) => { FileImpl::Sliced(parent_id, inner_rel_pos) => {
// Next time we don't need to check validity since // Next time we don't need to check validity since
// we have already done that for requesting URL if necessary // we have already done that for requesting URL if necessary.
self.fetch_blob_buf(sender, &parent_id, origin_in, return self.fetch_blob_buf(
rel_pos.slice_inner(&inner_rel_pos), false, response) sender,
&parent_id,
origin_in,
rel_pos.slice_inner(&inner_rel_pos),
false,
response,
);
},
} }
} }
}
fn try_fetch_file(&self, sender: &servo_channel::Sender<Data>, id: Uuid, check_url_validity: bool,
origin_in: FileOrigin, response: &mut Response)
-> Result<(), BlobURLStoreError> {
self.fetch_blob_buf(sender, &id, &origin_in, RelativePos::full_range(), check_url_validity, response)
}
fn dec_ref(&self, id: &Uuid, origin_in: &FileOrigin) -> Result<(), BlobURLStoreError> { fn dec_ref(&self, id: &Uuid, origin_in: &FileOrigin) -> Result<(), BlobURLStoreError> {
let (do_remove, opt_parent_id) = match self.entries.read().unwrap().get(id) { let (do_remove, opt_parent_id) = match self.entries.read().unwrap().get(id) {
@ -772,49 +773,62 @@ fn chunked_read(
} }
} }
fn chunked_fetch(sender: &servo_channel::Sender<Data>, fn chunked_fetch(
file: &mut File, size: usize, opt_filename: Option<String>, sender: servo_channel::Sender<Data>,
type_string: String, response: &mut Response, bytes: &mut Vec<u8>) { file: &mut File,
// First chunk response_body: ServoArc<Mutex<ResponseBody>>,
let mut buf = vec![0; CHUNK_SIZE]; ) {
match file.read(&mut buf) {
Ok(n) => {
buf.truncate(n);
let blob_buf = BlobBuf {
filename: opt_filename,
type_string: type_string,
size: size as u64,
bytes: buf,
};
bytes.extend_from_slice(&blob_buf.bytes);
let _ = sender.send(Data::Payload(blob_buf.bytes));
}
Err(_) => {
*response = Response::network_error(NetworkError::Internal("Opening file failed".into()));
return;
}
}
// Send the remaining chunks
loop { loop {
let mut buf = vec![0; CHUNK_SIZE]; let mut buf = vec![0; CHUNK_SIZE];
match file.read(&mut buf) { match file.read(&mut buf) {
Ok(0) => { Ok(0) | Err(_) => {
*response.body.lock().unwrap() = ResponseBody::Done(bytes.to_vec()); *response_body.lock().unwrap() = ResponseBody::Done(vec![]);
let _ = sender.send(Data::Done); let _ = sender.send(Data::Done);
return; return;
} },
Ok(n) => { Ok(n) => {
buf.truncate(n); buf.truncate(n);
let mut bytes = vec![];
bytes.extend_from_slice(&buf); bytes.extend_from_slice(&buf);
let _ = sender.send(Data::Payload(buf)); let _ = sender.send(Data::Payload(buf));
} },
Err(_) => {
*response = Response::network_error(NetworkError::Internal("Opening file failed".into()));
return;
}
} }
} }
} }
fn set_headers(headers: &mut HeaderMap, content_length: u64, mime: Mime, filename: Option<String>) {
headers.typed_insert(ContentLength(content_length));
headers.typed_insert(ContentType::from(mime.clone()));
let name = match filename {
Some(name) => name,
None => return,
};
let charset = mime.get_param(mime::CHARSET);
let charset = charset
.map(|c| c.as_ref().into())
.unwrap_or("us-ascii".to_owned());
// TODO(eijebong): Replace this once the typed header is there
// https://github.com/hyperium/headers/issues/8
headers.insert(
header::CONTENT_DISPOSITION,
HeaderValue::from_bytes(
format!(
"inline; {}",
if charset.to_lowercase() == "utf-8" {
format!(
"filename=\"{}\"",
String::from_utf8(name.as_bytes().into()).unwrap()
)
} else {
format!(
"filename*=\"{}\"''{}",
charset,
http_percent_encode(name.as_bytes())
)
}
)
.as_bytes(),
)
.unwrap(),
);
}