mirror of
https://github.com/servo/servo.git
synced 2025-08-04 13:10:20 +01:00
Finish asynchronous blob url fetching
This commit is contained in:
parent
67722d1943
commit
8538634210
3 changed files with 147 additions and 136 deletions
|
@ -2,20 +2,13 @@
|
||||||
* License, v. 2.0. If a copy of the MPL was not distributed with this
|
* License, v. 2.0. If a copy of the MPL was not distributed with this
|
||||||
* file, You can obtain one at https://mozilla.org/MPL/2.0/. */
|
* file, You can obtain one at https://mozilla.org/MPL/2.0/. */
|
||||||
|
|
||||||
|
use crate::fetch::methods::{Data, DoneChannel};
|
||||||
use crate::filemanager_thread::FileManager;
|
use crate::filemanager_thread::FileManager;
|
||||||
use fetch::methods::DoneChannel;
|
|
||||||
use headers_core::HeaderMapExt;
|
|
||||||
use headers_ext::{ContentLength, ContentType};
|
|
||||||
use http::header::{self, HeaderValue};
|
|
||||||
use http::HeaderMap;
|
|
||||||
use ipc_channel::ipc;
|
|
||||||
use mime::{self, Mime};
|
|
||||||
use net_traits::blob_url_store::parse_blob_url;
|
use net_traits::blob_url_store::parse_blob_url;
|
||||||
use net_traits::filemanager_thread::ReadFileProgress;
|
|
||||||
use net_traits::response::{Response, ResponseBody};
|
use net_traits::response::{Response, ResponseBody};
|
||||||
use net_traits::{http_percent_encode, NetworkError, ResourceFetchTiming};
|
use net_traits::{NetworkError, ResourceFetchTiming};
|
||||||
use servo_url::ServoUrl;
|
|
||||||
use servo_channel::channel;
|
use servo_channel::channel;
|
||||||
|
use servo_url::ServoUrl;
|
||||||
|
|
||||||
// TODO: Check on GET
|
// TODO: Check on GET
|
||||||
// https://w3c.github.io/FileAPI/#requestResponseModel
|
// https://w3c.github.io/FileAPI/#requestResponseModel
|
||||||
|
@ -24,13 +17,13 @@ use servo_channel::channel;
|
||||||
pub fn load_blob_async(
|
pub fn load_blob_async(
|
||||||
url: ServoUrl,
|
url: ServoUrl,
|
||||||
filemanager: FileManager,
|
filemanager: FileManager,
|
||||||
done_chan: &mut DoneChannel
|
done_chan: &mut DoneChannel,
|
||||||
) -> Response {
|
) -> Response {
|
||||||
let (id, origin) = match parse_blob_url(&url) {
|
let (id, origin) = match parse_blob_url(&url) {
|
||||||
Ok((id, origin)) => (id, origin),
|
Ok((id, origin)) => (id, origin),
|
||||||
Err(()) => {
|
Err(()) => {
|
||||||
return Response::network_error(NetworkError::Internal("Invalid blob url".into()));
|
return Response::network_error(NetworkError::Internal("Invalid blob url".into()));
|
||||||
}
|
},
|
||||||
};
|
};
|
||||||
|
|
||||||
let mut response = Response::new(url, ResourceFetchTiming::new(request.timing_type()));
|
let mut response = Response::new(url, ResourceFetchTiming::new(request.timing_type()));
|
||||||
|
@ -38,7 +31,11 @@ pub fn load_blob_async(
|
||||||
*done_chan = Some((sender.clone(), receiver));
|
*done_chan = Some((sender.clone(), receiver));
|
||||||
*response.body.lock().unwrap() = ResponseBody::Receiving(vec![]);
|
*response.body.lock().unwrap() = ResponseBody::Receiving(vec![]);
|
||||||
let check_url_validity = true;
|
let check_url_validity = true;
|
||||||
filemanager.fetch_file(sender, id, check_url_validity, origin, &mut response);
|
if let Err(err) = filemanager.fetch_file(&sender, id, check_url_validity, origin, &mut response)
|
||||||
|
{
|
||||||
|
let _ = sender.send(Data::Done);
|
||||||
|
return Response::network_error(NetworkError::Internal(err));
|
||||||
|
};
|
||||||
|
|
||||||
response
|
response
|
||||||
}
|
}
|
||||||
|
|
|
@ -657,7 +657,7 @@ fn scheme_fetch(
|
||||||
));
|
));
|
||||||
}
|
}
|
||||||
|
|
||||||
load_blob_async(url.clone(), context.filemanager.clone(), done_chan);
|
load_blob_async(url.clone(), context.filemanager.clone(), done_chan)
|
||||||
},
|
},
|
||||||
|
|
||||||
"ftp" => {
|
"ftp" => {
|
||||||
|
|
|
@ -2,22 +2,21 @@
|
||||||
* License, v. 2.0. If a copy of the MPL was not distributed with this
|
* License, v. 2.0. If a copy of the MPL was not distributed with this
|
||||||
* file, You can obtain one at https://mozilla.org/MPL/2.0/. */
|
* file, You can obtain one at https://mozilla.org/MPL/2.0/. */
|
||||||
|
|
||||||
|
use crate::fetch::methods::Data;
|
||||||
use embedder_traits::{EmbedderMsg, EmbedderProxy, FilterPattern};
|
use embedder_traits::{EmbedderMsg, EmbedderProxy, FilterPattern};
|
||||||
use ipc_channel::ipc::{self, IpcSender};
|
use headers_ext::{ContentLength, ContentType, HeaderMap, HeaderMapExt};
|
||||||
use fetch::methods::Data;
|
|
||||||
use headers_core::HeaderMapExt;
|
|
||||||
use headers_ext::{ContentLength, ContentType};
|
|
||||||
use http::HeaderMap;
|
|
||||||
use http::header::{self, HeaderValue};
|
use http::header::{self, HeaderValue};
|
||||||
|
use ipc_channel::ipc::{self, IpcSender};
|
||||||
use mime::{self, Mime};
|
use mime::{self, Mime};
|
||||||
use mime_guess::guess_mime_type_opt;
|
use mime_guess::guess_mime_type_opt;
|
||||||
use net_traits::{http_percent_encode, NetworkError};
|
|
||||||
use net_traits::blob_url_store::{BlobBuf, BlobURLStoreError};
|
use net_traits::blob_url_store::{BlobBuf, BlobURLStoreError};
|
||||||
use net_traits::filemanager_thread::{FileManagerResult, FileManagerThreadMsg, FileOrigin};
|
use net_traits::filemanager_thread::{FileManagerResult, FileManagerThreadMsg, FileOrigin};
|
||||||
use net_traits::filemanager_thread::{
|
use net_traits::filemanager_thread::{
|
||||||
FileManagerThreadError, ReadFileProgress, RelativePos, SelectedFile,
|
FileManagerThreadError, ReadFileProgress, RelativePos, SelectedFile,
|
||||||
};
|
};
|
||||||
|
use net_traits::http_percent_encode;
|
||||||
use net_traits::response::{Response, ResponseBody};
|
use net_traits::response::{Response, ResponseBody};
|
||||||
|
use servo_arc::Arc as ServoArc;
|
||||||
use servo_channel;
|
use servo_channel;
|
||||||
use servo_config::prefs::PREFS;
|
use servo_config::prefs::PREFS;
|
||||||
use std::collections::HashMap;
|
use std::collections::HashMap;
|
||||||
|
@ -26,7 +25,7 @@ use std::io::{Read, Seek, SeekFrom};
|
||||||
use std::ops::Index;
|
use std::ops::Index;
|
||||||
use std::path::{Path, PathBuf};
|
use std::path::{Path, PathBuf};
|
||||||
use std::sync::atomic::{self, AtomicBool, AtomicUsize, Ordering};
|
use std::sync::atomic::{self, AtomicBool, AtomicUsize, Ordering};
|
||||||
use std::sync::{Arc, RwLock};
|
use std::sync::{Arc, Mutex, RwLock};
|
||||||
use std::thread;
|
use std::thread;
|
||||||
use url::Url;
|
use url::Url;
|
||||||
use uuid::Uuid;
|
use uuid::Uuid;
|
||||||
|
@ -100,17 +99,27 @@ impl FileManager {
|
||||||
.expect("Thread spawning failed");
|
.expect("Thread spawning failed");
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn fetch_file(&self,
|
// Read a file for the Fetch implementation.
|
||||||
sender: servo_channel::Sender<Data>,
|
// It gets the required headers synchronously and reads the actual content
|
||||||
|
// in a separate thread.
|
||||||
|
pub fn fetch_file(
|
||||||
|
&self,
|
||||||
|
sender: &servo_channel::Sender<Data>,
|
||||||
id: Uuid,
|
id: Uuid,
|
||||||
check_url_validity: bool,
|
check_url_validity: bool,
|
||||||
origin: FileOrigin,
|
origin: FileOrigin,
|
||||||
response: &mut Response) {
|
response: &mut Response,
|
||||||
let store = self.store.clone();
|
) -> Result<(), String> {
|
||||||
let mut r2 = response.clone();
|
self.store
|
||||||
thread::Builder::new().name("read file".to_owned()).spawn(move || {
|
.fetch_blob_buf(
|
||||||
store.try_fetch_file(&sender, id, check_url_validity, origin, &mut r2)
|
sender,
|
||||||
}).expect("Thread spawning failed");
|
&id,
|
||||||
|
&origin,
|
||||||
|
RelativePos::full_range(),
|
||||||
|
check_url_validity,
|
||||||
|
response,
|
||||||
|
)
|
||||||
|
.map_err(|e| format!("{:?}", e))
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn promote_memory(
|
pub fn promote_memory(
|
||||||
|
@ -511,52 +520,36 @@ impl FileManagerStore {
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
fn fetch_blob_buf(&self, sender: &servo_channel::Sender<Data>,
|
fn fetch_blob_buf(
|
||||||
id: &Uuid, origin_in: &FileOrigin, rel_pos: RelativePos,
|
&self,
|
||||||
check_url_validity: bool, response: &mut Response) -> Result<(), BlobURLStoreError> {
|
sender: &servo_channel::Sender<Data>,
|
||||||
let mut bytes = vec![];
|
id: &Uuid,
|
||||||
|
origin_in: &FileOrigin,
|
||||||
|
rel_pos: RelativePos,
|
||||||
|
check_url_validity: bool,
|
||||||
|
response: &mut Response,
|
||||||
|
) -> Result<(), BlobURLStoreError> {
|
||||||
let file_impl = self.get_impl(id, origin_in, check_url_validity)?;
|
let file_impl = self.get_impl(id, origin_in, check_url_validity)?;
|
||||||
match file_impl {
|
match file_impl {
|
||||||
FileImpl::Memory(buf) => {
|
FileImpl::Memory(buf) => {
|
||||||
let range = rel_pos.to_abs_range(buf.size as usize);
|
let range = rel_pos.to_abs_range(buf.size as usize);
|
||||||
let blob_buf = BlobBuf {
|
let len = range.len() as u64;
|
||||||
filename: None,
|
|
||||||
type_string: buf.type_string,
|
|
||||||
size: range.len() as u64,
|
|
||||||
bytes: buf.bytes.index(range).to_vec(),
|
|
||||||
};
|
|
||||||
|
|
||||||
let content_type: Mime = blob_buf.type_string.parse().unwrap_or(mime::TEXT_PLAIN);
|
set_headers(
|
||||||
let charset = content_type.get_param(mime::CHARSET);
|
&mut response.headers,
|
||||||
let mut headers = HeaderMap::new();
|
len,
|
||||||
|
buf.type_string.parse().unwrap_or(mime::TEXT_PLAIN),
|
||||||
if let Some(name) = blob_buf.filename {
|
/* filename */ None,
|
||||||
let charset = charset.map(|c| c.as_ref().into()).unwrap_or("us-ascii".to_owned());
|
|
||||||
// TODO(eijebong): Replace this once the typed header is there
|
|
||||||
headers.insert(
|
|
||||||
header::CONTENT_DISPOSITION,
|
|
||||||
HeaderValue::from_bytes(
|
|
||||||
format!("inline; {}",
|
|
||||||
if charset.to_lowercase() == "utf-8" {
|
|
||||||
format!("filename=\"{}\"", String::from_utf8(name.as_bytes().into()).unwrap())
|
|
||||||
} else {
|
|
||||||
format!("filename*=\"{}\"''{}", charset, http_percent_encode(name.as_bytes()))
|
|
||||||
}
|
|
||||||
).as_bytes()
|
|
||||||
).unwrap()
|
|
||||||
);
|
);
|
||||||
}
|
|
||||||
|
|
||||||
headers.typed_insert(ContentLength(blob_buf.size as u64));
|
let mut bytes = vec![];
|
||||||
headers.typed_insert(ContentType::from(content_type.clone()));
|
bytes.extend_from_slice(buf.bytes.index(range));
|
||||||
|
|
||||||
bytes.extend_from_slice(&blob_buf.bytes);
|
let _ = sender.send(Data::Payload(bytes));
|
||||||
|
|
||||||
response.headers = headers;
|
|
||||||
*response.body.lock().unwrap() = ResponseBody::Done(bytes);
|
|
||||||
let _ = sender.send(Data::Done);
|
let _ = sender.send(Data::Done);
|
||||||
|
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
},
|
||||||
FileImpl::MetaDataOnly(metadata) => {
|
FileImpl::MetaDataOnly(metadata) => {
|
||||||
/* XXX: Snapshot state check (optional) https://w3c.github.io/FileAPI/#snapshot-state.
|
/* XXX: Snapshot state check (optional) https://w3c.github.io/FileAPI/#snapshot-state.
|
||||||
Concretely, here we create another file, and this file might not
|
Concretely, here we create another file, and this file might not
|
||||||
|
@ -564,45 +557,53 @@ impl FileManagerStore {
|
||||||
create_entry is called.
|
create_entry is called.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
let opt_filename = metadata.path.file_name()
|
let mut file = File::open(&metadata.path)
|
||||||
|
.map_err(|e| BlobURLStoreError::External(e.to_string()))?;
|
||||||
|
|
||||||
|
let range = rel_pos.to_abs_range(metadata.size as usize);
|
||||||
|
let range_start = range.start as u64;
|
||||||
|
let seeked_start = file
|
||||||
|
.seek(SeekFrom::Start(range_start))
|
||||||
|
.map_err(|e| BlobURLStoreError::External(e.to_string()))?;
|
||||||
|
if seeked_start != range_start {
|
||||||
|
return Err(BlobURLStoreError::InvalidEntry);
|
||||||
|
}
|
||||||
|
|
||||||
|
let filename = metadata
|
||||||
|
.path
|
||||||
|
.file_name()
|
||||||
.and_then(|osstr| osstr.to_str())
|
.and_then(|osstr| osstr.to_str())
|
||||||
.map(|s| s.to_string());
|
.map(|s| s.to_string());
|
||||||
|
|
||||||
let mime = guess_mime_type_opt(metadata.path.clone());
|
set_headers(
|
||||||
let range = rel_pos.to_abs_range(metadata.size as usize);
|
&mut response.headers,
|
||||||
|
metadata.size,
|
||||||
|
guess_mime_type_opt(metadata.path).unwrap_or(mime::TEXT_PLAIN),
|
||||||
|
filename,
|
||||||
|
);
|
||||||
|
|
||||||
let mut file = File::open(&metadata.path)
|
let body = response.body.clone();
|
||||||
.map_err(|e| BlobURLStoreError::External(e.to_string()))?;
|
let sender = sender.clone();
|
||||||
let seeked_start = file.seek(SeekFrom::Start(range.start as u64))
|
thread::Builder::new()
|
||||||
.map_err(|e| BlobURLStoreError::External(e.to_string()))?;
|
.name("fetch file".to_owned())
|
||||||
|
.spawn(move || chunked_fetch(sender, &mut file, body))
|
||||||
if seeked_start == (range.start as u64) {
|
.expect("Thread spawn failed");
|
||||||
let type_string = match mime {
|
|
||||||
Some(x) => format!("{}", x),
|
|
||||||
None => "".to_string(),
|
|
||||||
};
|
|
||||||
|
|
||||||
chunked_fetch(sender, &mut file, range.len(), opt_filename,
|
|
||||||
type_string, response, &mut bytes);
|
|
||||||
Ok(())
|
Ok(())
|
||||||
} else {
|
},
|
||||||
Err(BlobURLStoreError::InvalidEntry)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
FileImpl::Sliced(parent_id, inner_rel_pos) => {
|
FileImpl::Sliced(parent_id, inner_rel_pos) => {
|
||||||
// Next time we don't need to check validity since
|
// Next time we don't need to check validity since
|
||||||
// we have already done that for requesting URL if necessary
|
// we have already done that for requesting URL if necessary.
|
||||||
self.fetch_blob_buf(sender, &parent_id, origin_in,
|
return self.fetch_blob_buf(
|
||||||
rel_pos.slice_inner(&inner_rel_pos), false, response)
|
sender,
|
||||||
|
&parent_id,
|
||||||
|
origin_in,
|
||||||
|
rel_pos.slice_inner(&inner_rel_pos),
|
||||||
|
false,
|
||||||
|
response,
|
||||||
|
);
|
||||||
|
},
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
|
||||||
|
|
||||||
fn try_fetch_file(&self, sender: &servo_channel::Sender<Data>, id: Uuid, check_url_validity: bool,
|
|
||||||
origin_in: FileOrigin, response: &mut Response)
|
|
||||||
-> Result<(), BlobURLStoreError> {
|
|
||||||
self.fetch_blob_buf(sender, &id, &origin_in, RelativePos::full_range(), check_url_validity, response)
|
|
||||||
}
|
|
||||||
|
|
||||||
fn dec_ref(&self, id: &Uuid, origin_in: &FileOrigin) -> Result<(), BlobURLStoreError> {
|
fn dec_ref(&self, id: &Uuid, origin_in: &FileOrigin) -> Result<(), BlobURLStoreError> {
|
||||||
let (do_remove, opt_parent_id) = match self.entries.read().unwrap().get(id) {
|
let (do_remove, opt_parent_id) = match self.entries.read().unwrap().get(id) {
|
||||||
|
@ -772,49 +773,62 @@ fn chunked_read(
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn chunked_fetch(sender: &servo_channel::Sender<Data>,
|
fn chunked_fetch(
|
||||||
file: &mut File, size: usize, opt_filename: Option<String>,
|
sender: servo_channel::Sender<Data>,
|
||||||
type_string: String, response: &mut Response, bytes: &mut Vec<u8>) {
|
file: &mut File,
|
||||||
// First chunk
|
response_body: ServoArc<Mutex<ResponseBody>>,
|
||||||
let mut buf = vec![0; CHUNK_SIZE];
|
) {
|
||||||
match file.read(&mut buf) {
|
|
||||||
Ok(n) => {
|
|
||||||
buf.truncate(n);
|
|
||||||
let blob_buf = BlobBuf {
|
|
||||||
filename: opt_filename,
|
|
||||||
type_string: type_string,
|
|
||||||
size: size as u64,
|
|
||||||
bytes: buf,
|
|
||||||
};
|
|
||||||
|
|
||||||
bytes.extend_from_slice(&blob_buf.bytes);
|
|
||||||
|
|
||||||
let _ = sender.send(Data::Payload(blob_buf.bytes));
|
|
||||||
}
|
|
||||||
Err(_) => {
|
|
||||||
*response = Response::network_error(NetworkError::Internal("Opening file failed".into()));
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Send the remaining chunks
|
|
||||||
loop {
|
loop {
|
||||||
let mut buf = vec![0; CHUNK_SIZE];
|
let mut buf = vec![0; CHUNK_SIZE];
|
||||||
match file.read(&mut buf) {
|
match file.read(&mut buf) {
|
||||||
Ok(0) => {
|
Ok(0) | Err(_) => {
|
||||||
*response.body.lock().unwrap() = ResponseBody::Done(bytes.to_vec());
|
*response_body.lock().unwrap() = ResponseBody::Done(vec![]);
|
||||||
let _ = sender.send(Data::Done);
|
let _ = sender.send(Data::Done);
|
||||||
return;
|
return;
|
||||||
}
|
},
|
||||||
Ok(n) => {
|
Ok(n) => {
|
||||||
buf.truncate(n);
|
buf.truncate(n);
|
||||||
|
let mut bytes = vec![];
|
||||||
bytes.extend_from_slice(&buf);
|
bytes.extend_from_slice(&buf);
|
||||||
let _ = sender.send(Data::Payload(buf));
|
let _ = sender.send(Data::Payload(buf));
|
||||||
}
|
},
|
||||||
Err(_) => {
|
|
||||||
*response = Response::network_error(NetworkError::Internal("Opening file failed".into()));
|
|
||||||
return;
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fn set_headers(headers: &mut HeaderMap, content_length: u64, mime: Mime, filename: Option<String>) {
|
||||||
|
headers.typed_insert(ContentLength(content_length));
|
||||||
|
headers.typed_insert(ContentType::from(mime.clone()));
|
||||||
|
let name = match filename {
|
||||||
|
Some(name) => name,
|
||||||
|
None => return,
|
||||||
|
};
|
||||||
|
let charset = mime.get_param(mime::CHARSET);
|
||||||
|
let charset = charset
|
||||||
|
.map(|c| c.as_ref().into())
|
||||||
|
.unwrap_or("us-ascii".to_owned());
|
||||||
|
// TODO(eijebong): Replace this once the typed header is there
|
||||||
|
// https://github.com/hyperium/headers/issues/8
|
||||||
|
headers.insert(
|
||||||
|
header::CONTENT_DISPOSITION,
|
||||||
|
HeaderValue::from_bytes(
|
||||||
|
format!(
|
||||||
|
"inline; {}",
|
||||||
|
if charset.to_lowercase() == "utf-8" {
|
||||||
|
format!(
|
||||||
|
"filename=\"{}\"",
|
||||||
|
String::from_utf8(name.as_bytes().into()).unwrap()
|
||||||
|
)
|
||||||
|
} else {
|
||||||
|
format!(
|
||||||
|
"filename*=\"{}\"''{}",
|
||||||
|
charset,
|
||||||
|
http_percent_encode(name.as_bytes())
|
||||||
|
)
|
||||||
|
}
|
||||||
|
)
|
||||||
|
.as_bytes(),
|
||||||
|
)
|
||||||
|
.unwrap(),
|
||||||
|
);
|
||||||
}
|
}
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue