diff options
author | Kevin Cernekee <cernekee@chromium.org> | 2017-09-20 13:42:04 -0700 |
---|---|---|
committer | Samuel Thibault <samuel.thibault@ens-lyon.org> | 2017-09-24 20:04:09 +0200 |
commit | 3096116ff4a88acb3235b33b9c190996cd54b80c (patch) | |
tree | 85d216f50745122f954c169ad66284b0f2ac15cd | |
parent | 0b5a6fec4ed11326bcdc07c1ef5702df39f1bb45 (diff) | |
download | slirp-3096116ff4a88acb3235b33b9c190996cd54b80c.zip slirp-3096116ff4a88acb3235b33b9c190996cd54b80c.tar.gz slirp-3096116ff4a88acb3235b33b9c190996cd54b80c.tar.bz2 |
slirp: Fix intermittent send queue hangs on a socket
if_output() originally sent one mbuf per call and used the slirp->next_m
variable to keep track of where it left off. But nowadays it tries to
send all of the mbufs from the fastq, and one mbuf from each session on
the batchq. The next_m variable is both redundant and harmful: there is
a case[0] involving delayed packets in which next_m ends up pointing
to &slirp->if_batchq when an active session still exists, and this
blocks all traffic for that session until qemu is restarted.
The test case was created to reproduce a problem that was seen on
long-running Chromium OS VM tests[1] which rapidly create and
destroy ssh connections through hostfwd.
[0] https://pastebin.com/NNy6LreF
[1] https://bugs.chromium.org/p/chromium/issues/detail?id=766323
Signed-off-by: Kevin Cernekee <cernekee@chromium.org>
Signed-off-by: Samuel Thibault <samuel.thibault@ens-lyon.org>
-rw-r--r-- | if.c | 51 | ||||
-rw-r--r-- | slirp.h | 1 |
2 files changed, 17 insertions, 35 deletions
@@ -27,7 +27,6 @@ void if_init(Slirp *slirp) { slirp->if_fastq.qh_link = slirp->if_fastq.qh_rlink = &slirp->if_fastq; slirp->if_batchq.qh_link = slirp->if_batchq.qh_rlink = &slirp->if_batchq; - slirp->next_m = (struct mbuf *)&slirp->if_batchq; } /* @@ -96,10 +95,6 @@ void if_output(struct socket *so, struct mbuf *ifm) } } else { ifq = (struct mbuf *)slirp->if_batchq.qh_rlink; - /* Set next_m if the queue was empty so far */ - if ((struct quehead *)slirp->next_m == &slirp->if_batchq) { - slirp->next_m = ifm; - } } /* Create a new doubly linked list for this session */ @@ -138,21 +133,18 @@ diddit: } /* - * Send a packet - * We choose a packet based on its position in the output queues; + * Send one packet from each session. * If there are packets on the fastq, they are sent FIFO, before - * everything else. Otherwise we choose the first packet from the - * batchq and send it. the next packet chosen will be from the session - * after this one, then the session after that one, and so on.. So, - * for example, if there are 3 ftp session's fighting for bandwidth, + * everything else. Then we choose the first packet from each + * batchq session (socket) and send it. + * For example, if there are 3 ftp sessions fighting for bandwidth, * one packet will be sent from the first session, then one packet - * from the second session, then one packet from the third, then back - * to the first, etc. etc. + * from the second session, then one packet from the third. */ void if_start(Slirp *slirp) { uint64_t now = qemu_clock_get_ns(QEMU_CLOCK_REALTIME); - bool from_batchq, next_from_batchq; + bool from_batchq = false; struct mbuf *ifm, *ifm_next, *ifqt; DEBUG_CALL("if_start"); @@ -162,26 +154,29 @@ void if_start(Slirp *slirp) } slirp->if_start_busy = true; + struct mbuf *batch_head = NULL; + if (slirp->if_batchq.qh_link != &slirp->if_batchq) { + batch_head = (struct mbuf *)slirp->if_batchq.qh_link; + } + if (slirp->if_fastq.qh_link != &slirp->if_fastq) { ifm_next = (struct mbuf *)slirp->if_fastq.qh_link; - next_from_batchq = false; - } else if ((struct quehead *)slirp->next_m != &slirp->if_batchq) { - /* Nothing on fastq, pick up from batchq via next_m */ - ifm_next = slirp->next_m; - next_from_batchq = true; + } else if (batch_head) { + /* Nothing on fastq, pick up from batchq */ + ifm_next = batch_head; + from_batchq = true; } else { ifm_next = NULL; } while (ifm_next) { ifm = ifm_next; - from_batchq = next_from_batchq; ifm_next = ifm->ifq_next; if ((struct quehead *)ifm_next == &slirp->if_fastq) { /* No more packets in fastq, switch to batchq */ - ifm_next = slirp->next_m; - next_from_batchq = true; + ifm_next = batch_head; + from_batchq = true; } if ((struct quehead *)ifm_next == &slirp->if_batchq) { /* end of batchq */ @@ -194,11 +189,6 @@ void if_start(Slirp *slirp) continue; } - if (ifm == slirp->next_m) { - /* Set which packet to send on next iteration */ - slirp->next_m = ifm->ifq_next; - } - /* Remove it from the queue */ ifqt = ifm->ifq_prev; remque(ifm); @@ -209,15 +199,8 @@ void if_start(Slirp *slirp) insque(next, ifqt); ifs_remque(ifm); - if (!from_batchq) { - /* Next packet in fastq is from the same session */ ifm_next = next; - next_from_batchq = false; - } else if ((struct quehead *)slirp->next_m == &slirp->if_batchq) { - /* Set next_m and ifm_next if the session packet is now the - * only one on batchq */ - slirp->next_m = ifm_next = next; } } @@ -183,7 +183,6 @@ struct Slirp { /* if states */ struct quehead if_fastq; /* fast queue (for interactive data) */ struct quehead if_batchq; /* queue for non-interactive data */ - struct mbuf *next_m; /* pointer to next mbuf to output */ bool if_start_busy; /* avoid if_start recursion */ /* ip states */ |