mirror of
https://github.com/servo/servo.git
synced 2025-08-03 20:50:07 +01:00
Blob url's changes now build and test
This commit is contained in:
parent
e84e1d607b
commit
2f3affcfc8
3 changed files with 29 additions and 32 deletions
|
@ -13,9 +13,9 @@ use mime::{self, Mime};
|
|||
use net_traits::blob_url_store::parse_blob_url;
|
||||
use net_traits::filemanager_thread::ReadFileProgress;
|
||||
use net_traits::response::{Response, ResponseBody};
|
||||
use net_traits::{http_percent_encode, NetworkError};
|
||||
use net_traits::{http_percent_encode, NetworkError, ResourceFetchTiming};
|
||||
use servo_url::ServoUrl;
|
||||
use std::sync::mpsc::channel;
|
||||
use servo_channel::channel;
|
||||
|
||||
// TODO: Check on GET
|
||||
// https://w3c.github.io/FileAPI/#requestResponseModel
|
||||
|
@ -24,22 +24,21 @@ use std::sync::mpsc::channel;
|
|||
pub fn load_blob_async(
|
||||
url: ServoUrl,
|
||||
filemanager: FileManager,
|
||||
response: &Response,
|
||||
done_chan: &mut DoneChannel
|
||||
) -> Result<(), NetworkError> {
|
||||
)-> Response {
|
||||
let (id, origin) = match parse_blob_url(&url) {
|
||||
Ok((id, origin)) => (id, origin),
|
||||
Err(()) => {
|
||||
let e = format!("Invalid blob URL format {:?}", url);
|
||||
return Err(NetworkError::Internal(e));
|
||||
},
|
||||
return Response::network_error(NetworkError::Internal("Invalid blob url".into()));
|
||||
}
|
||||
};
|
||||
|
||||
let mut response = Response::new(url, ResourceFetchTiming::new(request.timing_type()));
|
||||
let (sender, receiver) = channel();
|
||||
*done_chan = Some((sender.clone(), receiver));
|
||||
*response.body.lock().unwrap() = ResponseBody::Receiving(vec![]);
|
||||
let check_url_validity = true;
|
||||
filemanager.fetch_file(sender, id, check_url_validity, origin, response);
|
||||
filemanager.fetch_file(sender, id, check_url_validity, origin, &mut response);
|
||||
|
||||
Ok(())
|
||||
response
|
||||
}
|
||||
|
|
|
@ -657,11 +657,7 @@ fn scheme_fetch(
|
|||
));
|
||||
}
|
||||
|
||||
let mut response = Response::new(url, ResourceFetchTiming::new(request.timing_type()));
|
||||
|
||||
load_blob_async(url.clone(), context.filemanager.clone(), &response, done_chan);
|
||||
|
||||
response
|
||||
load_blob_async(url.clone(), context.filemanager.clone(), done_chan);
|
||||
},
|
||||
|
||||
"ftp" => {
|
||||
|
|
|
@ -16,6 +16,7 @@ use net_traits::filemanager_thread::{
|
|||
FileManagerThreadError, ReadFileProgress, RelativePos, SelectedFile,
|
||||
};
|
||||
use net_traits::response::{Response, ResponseBody};
|
||||
use servo_channel;
|
||||
use servo_config::prefs::PREFS;
|
||||
use std::collections::HashMap;
|
||||
use std::fs::File;
|
||||
|
@ -23,7 +24,7 @@ use std::io::{Read, Seek, SeekFrom};
|
|||
use std::ops::Index;
|
||||
use std::path::{Path, PathBuf};
|
||||
use std::sync::atomic::{self, AtomicBool, AtomicUsize, Ordering};
|
||||
use std::sync::{Arc, RwLock, Mutex, mpsc};
|
||||
use std::sync::{Arc, RwLock};
|
||||
use std::thread;
|
||||
use url::Url;
|
||||
use uuid::Uuid;
|
||||
|
@ -98,15 +99,15 @@ impl FileManager {
|
|||
}
|
||||
|
||||
pub fn fetch_file(&self,
|
||||
sender: mpsc::Sender<Data>,
|
||||
sender: servo_channel::Sender<Data>,
|
||||
id: Uuid,
|
||||
check_url_validity: bool,
|
||||
origin: FileOrigin,
|
||||
response: &Response) {
|
||||
response: &mut Response) {
|
||||
let store = self.store.clone();
|
||||
let mut res_body = response.body.clone();
|
||||
let mut r2 = response.clone();
|
||||
thread::Builder::new().name("read file".to_owned()).spawn(move || {
|
||||
store.try_fetch_file(&sender, id, check_url_validity, origin, response, res_body)
|
||||
store.try_fetch_file(&sender, id, check_url_validity, origin, &mut r2)
|
||||
}).expect("Thread spawning failed");
|
||||
}
|
||||
|
||||
|
@ -508,9 +509,9 @@ impl FileManagerStore {
|
|||
)
|
||||
}
|
||||
|
||||
fn fetch_blob_buf(&self, sender: &mpsc::Sender<Data>,
|
||||
fn fetch_blob_buf(&self, sender: &servo_channel::Sender<Data>,
|
||||
id: &Uuid, origin_in: &FileOrigin, rel_pos: RelativePos,
|
||||
check_url_validity: bool, response: &Response, res_body: Arc<Mutex<ResponseBody>>) -> Result<(), BlobURLStoreError> {
|
||||
check_url_validity: bool, response: &mut Response) -> Result<(), BlobURLStoreError> {
|
||||
let mut bytes = vec![];
|
||||
let file_impl = self.get_impl(id, origin_in, check_url_validity)?;
|
||||
match file_impl {
|
||||
|
@ -544,7 +545,7 @@ impl FileManagerStore {
|
|||
bytes.extend_from_slice(&blob_buf.bytes);
|
||||
|
||||
response.headers = headers;
|
||||
*res_body.lock().unwrap() = ResponseBody::Done(bytes);
|
||||
*response.body.lock().unwrap() = ResponseBody::Done(bytes);
|
||||
let _ = sender.send(Data::Done);
|
||||
Ok(())
|
||||
}
|
||||
|
@ -574,7 +575,7 @@ impl FileManagerStore {
|
|||
};
|
||||
|
||||
chunked_fetch(sender, &mut file, range.len(), opt_filename,
|
||||
type_string, response, res_body, &mut bytes);
|
||||
type_string, response, &mut bytes);
|
||||
Ok(())
|
||||
} else {
|
||||
Err(BlobURLStoreError::InvalidEntry)
|
||||
|
@ -584,16 +585,15 @@ impl FileManagerStore {
|
|||
// Next time we don't need to check validity since
|
||||
// we have already done that for requesting URL if necessary
|
||||
self.fetch_blob_buf(sender, &parent_id, origin_in,
|
||||
rel_pos.slice_inner(&inner_rel_pos), false, response, res_body)
|
||||
rel_pos.slice_inner(&inner_rel_pos), false, response)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
fn try_fetch_file(&self, sender: &mpsc::Sender<Data>, id: Uuid, check_url_validity: bool,
|
||||
origin_in: FileOrigin, response: &Response, res_body: Arc<Mutex<ResponseBody>>)
|
||||
fn try_fetch_file(&self, sender: &servo_channel::Sender<Data>, id: Uuid, check_url_validity: bool,
|
||||
origin_in: FileOrigin, response: &mut Response)
|
||||
-> Result<(), BlobURLStoreError> {
|
||||
self.fetch_blob_buf(sender, &id, &origin_in, RelativePos::full_range(), check_url_validity,
|
||||
response, res_body)
|
||||
self.fetch_blob_buf(sender, &id, &origin_in, RelativePos::full_range(), check_url_validity, response)
|
||||
}
|
||||
|
||||
fn dec_ref(&self, id: &Uuid, origin_in: &FileOrigin) -> Result<(), BlobURLStoreError> {
|
||||
|
@ -764,9 +764,9 @@ fn chunked_read(
|
|||
}
|
||||
}
|
||||
|
||||
fn chunked_fetch(sender: &mpsc::Sender<Data>,
|
||||
fn chunked_fetch(sender: &servo_channel::Sender<Data>,
|
||||
file: &mut File, size: usize, opt_filename: Option<String>,
|
||||
type_string: String, response: &Response, res_body: Arc<Mutex<ResponseBody>>, bytes: &mut Vec<u8>) {
|
||||
type_string: String, response: &mut Response, bytes: &mut Vec<u8>) {
|
||||
// First chunk
|
||||
let mut buf = vec![0; CHUNK_SIZE];
|
||||
match file.read(&mut buf) {
|
||||
|
@ -778,7 +778,9 @@ fn chunked_fetch(sender: &mpsc::Sender<Data>,
|
|||
size: size as u64,
|
||||
bytes: buf,
|
||||
};
|
||||
|
||||
bytes.extend_from_slice(&blob_buf.bytes);
|
||||
|
||||
let _ = sender.send(Data::Payload(blob_buf.bytes));
|
||||
}
|
||||
Err(_) => {
|
||||
|
@ -792,7 +794,7 @@ fn chunked_fetch(sender: &mpsc::Sender<Data>,
|
|||
let mut buf = vec![0; CHUNK_SIZE];
|
||||
match file.read(&mut buf) {
|
||||
Ok(0) => {
|
||||
*res_body.lock().unwrap() = ResponseBody::Done(bytes.to_vec());
|
||||
*response.body.lock().unwrap() = ResponseBody::Done(bytes.to_vec());
|
||||
let _ = sender.send(Data::Done);
|
||||
return;
|
||||
}
|
||||
|
@ -801,7 +803,7 @@ fn chunked_fetch(sender: &mpsc::Sender<Data>,
|
|||
bytes.extend_from_slice(&buf);
|
||||
let _ = sender.send(Data::Payload(buf));
|
||||
}
|
||||
Err(e) => {
|
||||
Err(_) => {
|
||||
*response = Response::network_error(NetworkError::Internal("Opening file failed".into()));
|
||||
return;
|
||||
}
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue