aboutsummaryrefslogtreecommitdiffstats
path: root/components/net/filemanager_thread.rs
diff options
context:
space:
mode:
Diffstat (limited to 'components/net/filemanager_thread.rs')
-rw-r--r--components/net/filemanager_thread.rs38
1 files changed, 20 insertions, 18 deletions
diff --git a/components/net/filemanager_thread.rs b/components/net/filemanager_thread.rs
index 0be8de815b4..6d4fdb3f2d5 100644
--- a/components/net/filemanager_thread.rs
+++ b/components/net/filemanager_thread.rs
@@ -16,6 +16,7 @@ use net_traits::filemanager_thread::{
FileManagerThreadError, ReadFileProgress, RelativePos, SelectedFile,
};
use net_traits::response::{Response, ResponseBody};
+use servo_channel;
use servo_config::prefs::PREFS;
use std::collections::HashMap;
use std::fs::File;
@@ -23,7 +24,7 @@ use std::io::{Read, Seek, SeekFrom};
use std::ops::Index;
use std::path::{Path, PathBuf};
use std::sync::atomic::{self, AtomicBool, AtomicUsize, Ordering};
-use std::sync::{Arc, RwLock, Mutex, mpsc};
+use std::sync::{Arc, RwLock};
use std::thread;
use url::Url;
use uuid::Uuid;
@@ -98,15 +99,15 @@ impl FileManager {
}
pub fn fetch_file(&self,
- sender: mpsc::Sender<Data>,
+ sender: servo_channel::Sender<Data>,
id: Uuid,
check_url_validity: bool,
origin: FileOrigin,
- response: &Response) {
+ response: &mut Response) {
let store = self.store.clone();
- let mut res_body = response.body.clone();
+ let mut r2 = response.clone();
thread::Builder::new().name("read file".to_owned()).spawn(move || {
- store.try_fetch_file(&sender, id, check_url_validity, origin, response, res_body)
+ store.try_fetch_file(&sender, id, check_url_validity, origin, &mut r2)
}).expect("Thread spawning failed");
}
@@ -508,9 +509,9 @@ impl FileManagerStore {
)
}
- fn fetch_blob_buf(&self, sender: &mpsc::Sender<Data>,
+ fn fetch_blob_buf(&self, sender: &servo_channel::Sender<Data>,
id: &Uuid, origin_in: &FileOrigin, rel_pos: RelativePos,
- check_url_validity: bool, response: &Response, res_body: Arc<Mutex<ResponseBody>>) -> Result<(), BlobURLStoreError> {
+ check_url_validity: bool, response: &mut Response) -> Result<(), BlobURLStoreError> {
let mut bytes = vec![];
let file_impl = self.get_impl(id, origin_in, check_url_validity)?;
match file_impl {
@@ -544,7 +545,7 @@ impl FileManagerStore {
bytes.extend_from_slice(&blob_buf.bytes);
response.headers = headers;
- *res_body.lock().unwrap() = ResponseBody::Done(bytes);
+ *response.body.lock().unwrap() = ResponseBody::Done(bytes);
let _ = sender.send(Data::Done);
Ok(())
}
@@ -574,7 +575,7 @@ impl FileManagerStore {
};
chunked_fetch(sender, &mut file, range.len(), opt_filename,
- type_string, response, res_body, &mut bytes);
+ type_string, response, &mut bytes);
Ok(())
} else {
Err(BlobURLStoreError::InvalidEntry)
@@ -584,16 +585,15 @@ impl FileManagerStore {
// Next time we don't need to check validity since
// we have already done that for requesting URL if necessary
self.fetch_blob_buf(sender, &parent_id, origin_in,
- rel_pos.slice_inner(&inner_rel_pos), false, response, res_body)
+ rel_pos.slice_inner(&inner_rel_pos), false, response)
}
}
}
- fn try_fetch_file(&self, sender: &mpsc::Sender<Data>, id: Uuid, check_url_validity: bool,
- origin_in: FileOrigin, response: &Response, res_body: Arc<Mutex<ResponseBody>>)
+ fn try_fetch_file(&self, sender: &servo_channel::Sender<Data>, id: Uuid, check_url_validity: bool,
+ origin_in: FileOrigin, response: &mut Response)
-> Result<(), BlobURLStoreError> {
- self.fetch_blob_buf(sender, &id, &origin_in, RelativePos::full_range(), check_url_validity,
- response, res_body)
+ self.fetch_blob_buf(sender, &id, &origin_in, RelativePos::full_range(), check_url_validity, response)
}
fn dec_ref(&self, id: &Uuid, origin_in: &FileOrigin) -> Result<(), BlobURLStoreError> {
@@ -764,9 +764,9 @@ fn chunked_read(
}
}
-fn chunked_fetch(sender: &mpsc::Sender<Data>,
+fn chunked_fetch(sender: &servo_channel::Sender<Data>,
file: &mut File, size: usize, opt_filename: Option<String>,
- type_string: String, response: &Response, res_body: Arc<Mutex<ResponseBody>>, bytes: &mut Vec<u8>) {
+ type_string: String, response: &mut Response, bytes: &mut Vec<u8>) {
// First chunk
let mut buf = vec![0; CHUNK_SIZE];
match file.read(&mut buf) {
@@ -778,7 +778,9 @@ fn chunked_fetch(sender: &mpsc::Sender<Data>,
size: size as u64,
bytes: buf,
};
+
bytes.extend_from_slice(&blob_buf.bytes);
+
let _ = sender.send(Data::Payload(blob_buf.bytes));
}
Err(_) => {
@@ -792,7 +794,7 @@ fn chunked_fetch(sender: &mpsc::Sender<Data>,
let mut buf = vec![0; CHUNK_SIZE];
match file.read(&mut buf) {
Ok(0) => {
- *res_body.lock().unwrap() = ResponseBody::Done(bytes.to_vec());
+ *response.body.lock().unwrap() = ResponseBody::Done(bytes.to_vec());
let _ = sender.send(Data::Done);
return;
}
@@ -801,7 +803,7 @@ fn chunked_fetch(sender: &mpsc::Sender<Data>,
bytes.extend_from_slice(&buf);
let _ = sender.send(Data::Payload(buf));
}
- Err(e) => {
+ Err(_) => {
*response = Response::network_error(NetworkError::Internal("Opening file failed".into()));
return;
}