@@ -221,10 +221,26 @@ where C: Connect,
|
|||||||
e.into()
|
e.into()
|
||||||
});
|
});
|
||||||
|
|
||||||
let resp = race.and_then(move |mut client| {
|
let resp = race.and_then(move |client| {
|
||||||
|
use proto::dispatch::ClientMsg;
|
||||||
|
|
||||||
let (callback, rx) = oneshot::channel();
|
let (callback, rx) = oneshot::channel();
|
||||||
client.tx.borrow_mut().start_send(proto::dispatch::ClientMsg::Request(head, body, callback)).unwrap();
|
|
||||||
client.should_close = false;
|
match client.tx.borrow_mut().start_send(ClientMsg::Request(head, body, callback)) {
|
||||||
|
Ok(_) => (),
|
||||||
|
Err(e) => match e.into_inner() {
|
||||||
|
ClientMsg::Request(_, _, callback) => {
|
||||||
|
error!("pooled connection was not ready, this is a hyper bug");
|
||||||
|
let err = io::Error::new(
|
||||||
|
io::ErrorKind::BrokenPipe,
|
||||||
|
"pool selected dead connection",
|
||||||
|
);
|
||||||
|
let _ = callback.send(Err(::Error::Io(err)));
|
||||||
|
},
|
||||||
|
_ => unreachable!("ClientMsg::Request was just sent"),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
rx.then(|res| {
|
rx.then(|res| {
|
||||||
match res {
|
match res {
|
||||||
Ok(Ok(res)) => Ok(res),
|
Ok(Ok(res)) => Ok(res),
|
||||||
@@ -256,8 +272,8 @@ impl<C, B> fmt::Debug for Client<C, B> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
struct HyperClient<B> {
|
struct HyperClient<B> {
|
||||||
tx: RefCell<::futures::sync::mpsc::Sender<proto::dispatch::ClientMsg<B>>>,
|
|
||||||
should_close: bool,
|
should_close: bool,
|
||||||
|
tx: RefCell<::futures::sync::mpsc::Sender<proto::dispatch::ClientMsg<B>>>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<B> Clone for HyperClient<B> {
|
impl<B> Clone for HyperClient<B> {
|
||||||
@@ -269,6 +285,15 @@ impl<B> Clone for HyperClient<B> {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
impl<B> self::pool::Ready for HyperClient<B> {
|
||||||
|
fn poll_ready(&mut self) -> Poll<(), ()> {
|
||||||
|
self.tx
|
||||||
|
.borrow_mut()
|
||||||
|
.poll_ready()
|
||||||
|
.map_err(|_| ())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
impl<B> Drop for HyperClient<B> {
|
impl<B> Drop for HyperClient<B> {
|
||||||
fn drop(&mut self) {
|
fn drop(&mut self) {
|
||||||
if self.should_close {
|
if self.should_close {
|
||||||
@@ -497,3 +522,4 @@ mod background {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -15,6 +15,15 @@ pub struct Pool<T> {
|
|||||||
inner: Rc<RefCell<PoolInner<T>>>,
|
inner: Rc<RefCell<PoolInner<T>>>,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Before using a pooled connection, make sure the sender is not dead.
|
||||||
|
//
|
||||||
|
// This is a trait to allow the `client::pool::tests` to work for `i32`.
|
||||||
|
//
|
||||||
|
// See https://github.com/hyperium/hyper/issues/1429
|
||||||
|
pub trait Ready {
|
||||||
|
fn poll_ready(&mut self) -> Poll<(), ()>;
|
||||||
|
}
|
||||||
|
|
||||||
struct PoolInner<T> {
|
struct PoolInner<T> {
|
||||||
enabled: bool,
|
enabled: bool,
|
||||||
// These are internal Conns sitting in the event loop in the KeepAlive
|
// These are internal Conns sitting in the event loop in the KeepAlive
|
||||||
@@ -256,7 +265,7 @@ pub struct Checkout<T> {
|
|||||||
parked: Option<relay::Receiver<Entry<T>>>,
|
parked: Option<relay::Receiver<Entry<T>>>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<T: Clone> Future for Checkout<T> {
|
impl<T: Ready + Clone> Future for Checkout<T> {
|
||||||
type Item = Pooled<T>;
|
type Item = Pooled<T>;
|
||||||
type Error = io::Error;
|
type Error = io::Error;
|
||||||
|
|
||||||
@@ -282,21 +291,22 @@ impl<T: Clone> Future for Checkout<T> {
|
|||||||
let mut should_remove = false;
|
let mut should_remove = false;
|
||||||
let entry = self.pool.inner.borrow_mut().idle.get_mut(key).and_then(|list| {
|
let entry = self.pool.inner.borrow_mut().idle.get_mut(key).and_then(|list| {
|
||||||
trace!("Checkout::poll key found {:?}", key);
|
trace!("Checkout::poll key found {:?}", key);
|
||||||
while let Some(entry) = list.pop() {
|
while let Some(mut entry) = list.pop() {
|
||||||
match entry.status.get() {
|
match entry.status.get() {
|
||||||
TimedKA::Idle(idle_at) if !expiration.expires(idle_at) => {
|
TimedKA::Idle(idle_at) if !expiration.expires(idle_at) => {
|
||||||
debug!("found idle connection for {:?}", key);
|
if let Ok(Async::Ready(())) = entry.value.poll_ready() {
|
||||||
should_remove = list.is_empty();
|
debug!("found idle connection for {:?}", key);
|
||||||
return Some(entry);
|
should_remove = list.is_empty();
|
||||||
|
return Some(entry);
|
||||||
|
}
|
||||||
},
|
},
|
||||||
_ => {
|
_ => {},
|
||||||
trace!("Checkout::poll removing unacceptable pooled {:?}", key);
|
|
||||||
// every other case the Entry should just be dropped
|
|
||||||
// 1. Idle but expired
|
|
||||||
// 2. Busy (something else somehow took it?)
|
|
||||||
// 3. Disabled don't reuse of course
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
trace!("Checkout::poll removing unacceptable pooled {:?}", key);
|
||||||
|
// every other case the Entry should just be dropped
|
||||||
|
// 1. Idle but expired
|
||||||
|
// 2. Busy (something else somehow took it?)
|
||||||
|
// 3. Disabled don't reuse of course
|
||||||
}
|
}
|
||||||
should_remove = true;
|
should_remove = true;
|
||||||
None
|
None
|
||||||
@@ -347,10 +357,16 @@ impl Expiration {
|
|||||||
mod tests {
|
mod tests {
|
||||||
use std::rc::Rc;
|
use std::rc::Rc;
|
||||||
use std::time::Duration;
|
use std::time::Duration;
|
||||||
use futures::{Async, Future};
|
use futures::{Async, Future, Poll};
|
||||||
use futures::future;
|
use futures::future;
|
||||||
use proto::KeepAlive;
|
use proto::KeepAlive;
|
||||||
use super::Pool;
|
use super::{Ready, Pool};
|
||||||
|
|
||||||
|
impl Ready for i32 {
|
||||||
|
fn poll_ready(&mut self) -> Poll<(), ()> {
|
||||||
|
Ok(Async::Ready(()))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_pool_checkout_smoke() {
|
fn test_pool_checkout_smoke() {
|
||||||
|
|||||||
@@ -954,6 +954,60 @@ mod dispatch_impl {
|
|||||||
assert_eq!(closes.load(Ordering::Relaxed), 1);
|
assert_eq!(closes.load(Ordering::Relaxed), 1);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn conn_drop_prevents_pool_checkout() {
|
||||||
|
// a drop might happen for any sort of reason, and we can protect
|
||||||
|
// against a lot of them, but if the `Core` is dropped, we can't
|
||||||
|
// really catch that. So, this is case to always check.
|
||||||
|
//
|
||||||
|
// See https://github.com/hyperium/hyper/issues/1429
|
||||||
|
|
||||||
|
use std::error::Error;
|
||||||
|
let _ = pretty_env_logger::try_init();
|
||||||
|
|
||||||
|
let server = TcpListener::bind("127.0.0.1:0").unwrap();
|
||||||
|
let addr = server.local_addr().unwrap();
|
||||||
|
let mut core = Core::new().unwrap();
|
||||||
|
let handle = core.handle();
|
||||||
|
|
||||||
|
let (tx1, rx1) = oneshot::channel();
|
||||||
|
|
||||||
|
thread::spawn(move || {
|
||||||
|
let mut sock = server.accept().unwrap().0;
|
||||||
|
sock.set_read_timeout(Some(Duration::from_secs(5))).unwrap();
|
||||||
|
sock.set_write_timeout(Some(Duration::from_secs(5))).unwrap();
|
||||||
|
let mut buf = [0; 4096];
|
||||||
|
sock.read(&mut buf).expect("read 1");
|
||||||
|
sock.write_all(b"HTTP/1.1 200 OK\r\nContent-Length: 0\r\n\r\n").unwrap();
|
||||||
|
sock.read(&mut buf).expect("read 2");
|
||||||
|
let _ = tx1.send(());
|
||||||
|
});
|
||||||
|
|
||||||
|
let uri = format!("http://{}/a", addr).parse::<hyper::Uri>().unwrap();
|
||||||
|
|
||||||
|
let client = Client::new(&handle);
|
||||||
|
let res = client.get(uri.clone()).and_then(move |res| {
|
||||||
|
assert_eq!(res.status(), hyper::StatusCode::Ok);
|
||||||
|
res.body().concat2()
|
||||||
|
});
|
||||||
|
|
||||||
|
core.run(res).unwrap();
|
||||||
|
|
||||||
|
// drop previous Core
|
||||||
|
core = Core::new().unwrap();
|
||||||
|
let timeout = Timeout::new(Duration::from_millis(200), &core.handle()).unwrap();
|
||||||
|
let rx = rx1.map_err(|_| hyper::Error::Io(io::Error::new(io::ErrorKind::Other, "thread panicked")));
|
||||||
|
let rx = rx.and_then(move |_| timeout.map_err(|e| e.into()));
|
||||||
|
|
||||||
|
let res = client.get(uri);
|
||||||
|
// this does trigger an 'event loop gone' error, but before, it would
|
||||||
|
// panic internally on a `SendError`, which is what we're testing against.
|
||||||
|
let err = core.run(res.join(rx).map(|r| r.0)).unwrap_err();
|
||||||
|
assert_eq!(err.description(), "event loop gone");
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn client_custom_executor() {
|
fn client_custom_executor() {
|
||||||
let server = TcpListener::bind("127.0.0.1:0").unwrap();
|
let server = TcpListener::bind("127.0.0.1:0").unwrap();
|
||||||
|
|||||||
Reference in New Issue
Block a user