/*
 * Copyright (c) 1995 Danny Gasparovski.
 *
 * Please read the file COPYRIGHT for the
 * terms and conditions of the copyright.
 */

#include <slirp.h>
#include "qemu/timer.h"

static void
ifs_insque(struct mbuf *ifm, struct mbuf *ifmhead)
{
        ifm->ifs_next = ifmhead->ifs_next;
        ifmhead->ifs_next = ifm;
        ifm->ifs_prev = ifmhead;
        ifm->ifs_next->ifs_prev = ifm;
}

static void
ifs_remque(struct mbuf *ifm)
{
        ifm->ifs_prev->ifs_next = ifm->ifs_next;
        ifm->ifs_next->ifs_prev = ifm->ifs_prev;
}

void
if_init(Slirp *slirp)
{
    slirp->if_fastq.ifq_next = slirp->if_fastq.ifq_prev = &slirp->if_fastq;
    slirp->if_batchq.ifq_next = slirp->if_batchq.ifq_prev = &slirp->if_batchq;
    slirp->next_m = &slirp->if_batchq;
}

/*
 * if_output: Queue packet into an output queue.
 * There are 2 output queue's, if_fastq and if_batchq.
 * Each output queue is a doubly linked list of double linked lists
 * of mbufs, each list belonging to one "session" (socket).  This
 * way, we can output packets fairly by sending one packet from each
 * session, instead of all the packets from one session, then all packets
 * from the next session, etc.  Packets on the if_fastq get absolute
 * priority, but if one session hogs the link, it gets "downgraded"
 * to the batchq until it runs out of packets, then it'll return
 * to the fastq (eg. if the user does an ls -alR in a telnet session,
 * it'll temporarily get downgraded to the batchq)
 */
void
if_output(struct socket *so, struct mbuf *ifm)
{
        Slirp *slirp = ifm->slirp;
        struct mbuf *ifq;
        int on_fastq = 1;

        DEBUG_CALL("if_output");
        DEBUG_ARG("so = %lx", (long)so);
        DEBUG_ARG("ifm = %lx", (long)ifm);

        /*
         * First remove the mbuf from m_usedlist,
         * since we're gonna use m_next and m_prev ourselves
         * XXX Shouldn't need this, gotta change dtom() etc.
         */
        if (ifm->m_flags & M_USEDLIST) {
                remque(ifm);
                ifm->m_flags &= ~M_USEDLIST;
        }

        /*
         * See if there's already a batchq list for this session.
         * This can include an interactive session, which should go on fastq,
         * but gets too greedy... hence it'll be downgraded from fastq to batchq.
         * We mustn't put this packet back on the fastq (or we'll send it out of order)
         * XXX add cache here?
         */
        for (ifq = slirp->if_batchq.ifq_prev; ifq && ifq != &slirp->if_batchq;
             ifq = ifq->ifq_prev) {
                if (so == ifq->ifq_so) {
                        /* A match! */
                        ifm->ifq_so = so;
                        ifs_insque(ifm, ifq->ifs_prev);
                        goto diddit;
                }
        }

        /* No match, check which queue to put it on */
        if (so && (so->so_iptos & IPTOS_LOWDELAY)) {
                ifq = slirp->if_fastq.ifq_prev;
                on_fastq = 1;
                /*
                 * Check if this packet is a part of the last
                 * packet's session
                 */
                if (ifq->ifq_so == so) {
                        ifm->ifq_so = so;
                        ifs_insque(ifm, ifq->ifs_prev);
                        goto diddit;
                }
        } else {
                ifq = slirp->if_batchq.ifq_prev;
                /* Set next_m if the queue was empty so far */
                if (slirp->next_m == &slirp->if_batchq) {
                    slirp->next_m = ifm;
                }
        }

        /* Create a new doubly linked list for this session */
        ifm->ifq_so = so;
        ifs_init(ifm);
        insque(ifm, ifq);

diddit:
        if (so) {
                /* Update *_queued */
                so->so_queued++;
                so->so_nqueued++;
                /*
                 * Check if the interactive session should be downgraded to
                 * the batchq.  A session is downgraded if it has queued 6
                 * packets without pausing, and at least 3 of those packets
                 * have been sent over the link
                 * (XXX These are arbitrary numbers, probably not optimal..)
                 */
                if (on_fastq && ((so->so_nqueued >= 6) &&
                                 (so->so_nqueued - so->so_queued) >= 3)) {

                        /* Remove from current queue... */
                        remque(ifm->ifs_next);

                        /* ...And insert in the new.  That'll teach ya! */
                        insque(ifm->ifs_next, &slirp->if_batchq);
                }
        }

#ifndef FULL_BOLT
        /*
         * This prevents us from malloc()ing too many mbufs
         */
        if_start(ifm->slirp);
#endif
}

/*
 * Send a packet
 * We choose a packet based on its position in the output queues;
 * If there are packets on the fastq, they are sent FIFO, before
 * everything else.  Otherwise we choose the first packet from the
 * batchq and send it.  the next packet chosen will be from the session
 * after this one, then the session after that one, and so on..  So,
 * for example, if there are 3 ftp session's fighting for bandwidth,
 * one packet will be sent from the first session, then one packet
 * from the second session, then one packet from the third, then back
 * to the first, etc. etc.
 */
void if_start(Slirp *slirp)
{
    uint64_t now = qemu_clock_get_ns(QEMU_CLOCK_REALTIME);
    bool from_batchq, next_from_batchq;
    struct mbuf *ifm, *ifm_next, *ifqt;

    DEBUG_CALL("if_start");

    if (slirp->if_start_busy) {
        return;
    }
    slirp->if_start_busy = true;

    if (slirp->if_fastq.ifq_next != &slirp->if_fastq) {
        ifm_next = slirp->if_fastq.ifq_next;
        next_from_batchq = false;
    } else if (slirp->next_m != &slirp->if_batchq) {
        /* Nothing on fastq, pick up from batchq via next_m */
        ifm_next = slirp->next_m;
        next_from_batchq = true;
    } else {
        ifm_next = NULL;
    }

    while (ifm_next) {
        ifm = ifm_next;
        from_batchq = next_from_batchq;

        ifm_next = ifm->ifq_next;
        if (ifm_next == &slirp->if_fastq) {
            /* No more packets in fastq, switch to batchq */
            ifm_next = slirp->next_m;
            next_from_batchq = true;
        }
        if (ifm_next == &slirp->if_batchq) {
            /* end of batchq */
            ifm_next = NULL;
        }

        /* Try to send packet unless it already expired */
        if (ifm->expiration_date >= now && !if_encap(slirp, ifm)) {
            /* Packet is delayed due to pending ARP resolution */
            continue;
        }

        if (ifm == slirp->next_m) {
            /* Set which packet to send on next iteration */
            slirp->next_m = ifm->ifq_next;
        }

        /* Remove it from the queue */
        ifqt = ifm->ifq_prev;
        remque(ifm);

        /* If there are more packets for this session, re-queue them */
        if (ifm->ifs_next != ifm) {
            struct mbuf *next = ifm->ifs_next;

            insque(next, ifqt);
            ifs_remque(ifm);

            if (!from_batchq) {
                /* Next packet in fastq is from the same session */
                ifm_next = next;
                next_from_batchq = false;
            } else if (slirp->next_m == &slirp->if_batchq) {
                /* Set next_m and ifm_next if the session packet is now the
                 * only one on batchq */
                slirp->next_m = ifm_next = next;
            }
        }

        /* Update so_queued */
        if (ifm->ifq_so && --ifm->ifq_so->so_queued == 0) {
            /* If there's no more queued, reset nqueued */
            ifm->ifq_so->so_nqueued = 0;
        }

        m_free(ifm);
    }

    slirp->if_start_busy = false;
}