mirror of
https://git.hardenedbsd.org/hardenedbsd/HardenedBSD.git
synced 2024-11-22 11:14:18 +01:00
29363fb446
Remove ancient SCCS tags from the tree, automated scripting, with two minor fixup to keep things compiling. All the common forms in the tree were removed with a perl script. Sponsored by: Netflix
425 lines
11 KiB
C
425 lines
11 KiB
C
/*-
|
|
* SPDX-License-Identifier: BSD-3-Clause
|
|
*
|
|
* Copyright (c) 1982, 1986, 1989, 1993
|
|
* The Regents of the University of California. All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
* 3. Neither the name of the University nor the names of its contributors
|
|
* may be used to endorse or promote products derived from this software
|
|
* without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*/
|
|
|
|
#ifndef _NET_IFQ_H_
|
|
#define _NET_IFQ_H_
|
|
|
|
#ifdef _KERNEL
|
|
#include <sys/mbuf.h> /* ifqueue only? */
|
|
#include <sys/buf_ring.h>
|
|
#include <net/vnet.h>
|
|
#endif /* _KERNEL */
|
|
#include <sys/lock.h> /* XXX */
|
|
#include <sys/mutex.h> /* struct ifqueue */
|
|
|
|
/*
|
|
* Couple of ugly extra definitions that are required since ifq.h
|
|
* is splitted from if_var.h.
|
|
*/
|
|
#define IF_DUNIT_NONE -1
|
|
|
|
#include <net/altq/if_altq.h>
|
|
|
|
/*
|
|
* Structure defining a queue for a network interface.
|
|
*/
|
|
struct ifqueue {
|
|
struct mbuf *ifq_head;
|
|
struct mbuf *ifq_tail;
|
|
int ifq_len;
|
|
int ifq_maxlen;
|
|
struct mtx ifq_mtx;
|
|
};
|
|
|
|
#ifdef _KERNEL
|
|
/*
|
|
* Output queues (ifp->if_snd) and slow device input queues (*ifp->if_slowq)
|
|
* are queues of messages stored on ifqueue structures
|
|
* (defined above). Entries are added to and deleted from these structures
|
|
* by these macros.
|
|
*/
|
|
#define IF_LOCK(ifq) mtx_lock(&(ifq)->ifq_mtx)
|
|
#define IF_UNLOCK(ifq) mtx_unlock(&(ifq)->ifq_mtx)
|
|
#define IF_LOCK_ASSERT(ifq) mtx_assert(&(ifq)->ifq_mtx, MA_OWNED)
|
|
#define _IF_QFULL(ifq) ((ifq)->ifq_len >= (ifq)->ifq_maxlen)
|
|
#define _IF_QLEN(ifq) ((ifq)->ifq_len)
|
|
|
|
#define _IF_ENQUEUE(ifq, m) do { \
|
|
(m)->m_nextpkt = NULL; \
|
|
if ((ifq)->ifq_tail == NULL) \
|
|
(ifq)->ifq_head = m; \
|
|
else \
|
|
(ifq)->ifq_tail->m_nextpkt = m; \
|
|
(ifq)->ifq_tail = m; \
|
|
(ifq)->ifq_len++; \
|
|
} while (0)
|
|
|
|
#define IF_ENQUEUE(ifq, m) do { \
|
|
IF_LOCK(ifq); \
|
|
_IF_ENQUEUE(ifq, m); \
|
|
IF_UNLOCK(ifq); \
|
|
} while (0)
|
|
|
|
#define _IF_PREPEND(ifq, m) do { \
|
|
(m)->m_nextpkt = (ifq)->ifq_head; \
|
|
if ((ifq)->ifq_tail == NULL) \
|
|
(ifq)->ifq_tail = (m); \
|
|
(ifq)->ifq_head = (m); \
|
|
(ifq)->ifq_len++; \
|
|
} while (0)
|
|
|
|
#define IF_PREPEND(ifq, m) do { \
|
|
IF_LOCK(ifq); \
|
|
_IF_PREPEND(ifq, m); \
|
|
IF_UNLOCK(ifq); \
|
|
} while (0)
|
|
|
|
#define _IF_DEQUEUE(ifq, m) do { \
|
|
(m) = (ifq)->ifq_head; \
|
|
if (m) { \
|
|
if (((ifq)->ifq_head = (m)->m_nextpkt) == NULL) \
|
|
(ifq)->ifq_tail = NULL; \
|
|
(m)->m_nextpkt = NULL; \
|
|
(ifq)->ifq_len--; \
|
|
} \
|
|
} while (0)
|
|
|
|
#define IF_DEQUEUE(ifq, m) do { \
|
|
IF_LOCK(ifq); \
|
|
_IF_DEQUEUE(ifq, m); \
|
|
IF_UNLOCK(ifq); \
|
|
} while (0)
|
|
|
|
#define _IF_DEQUEUE_ALL(ifq, m) do { \
|
|
(m) = (ifq)->ifq_head; \
|
|
(ifq)->ifq_head = (ifq)->ifq_tail = NULL; \
|
|
(ifq)->ifq_len = 0; \
|
|
} while (0)
|
|
|
|
#define IF_DEQUEUE_ALL(ifq, m) do { \
|
|
IF_LOCK(ifq); \
|
|
_IF_DEQUEUE_ALL(ifq, m); \
|
|
IF_UNLOCK(ifq); \
|
|
} while (0)
|
|
|
|
#define _IF_POLL(ifq, m) ((m) = (ifq)->ifq_head)
|
|
#define IF_POLL(ifq, m) _IF_POLL(ifq, m)
|
|
|
|
#define _IF_DRAIN(ifq) do { \
|
|
struct mbuf *m; \
|
|
for (;;) { \
|
|
_IF_DEQUEUE(ifq, m); \
|
|
if (m == NULL) \
|
|
break; \
|
|
m_freem(m); \
|
|
} \
|
|
} while (0)
|
|
|
|
#define IF_DRAIN(ifq) do { \
|
|
IF_LOCK(ifq); \
|
|
_IF_DRAIN(ifq); \
|
|
IF_UNLOCK(ifq); \
|
|
} while(0)
|
|
|
|
int if_handoff(struct ifqueue *ifq, struct mbuf *m, struct ifnet *ifp,
|
|
int adjust);
|
|
#define IF_HANDOFF(ifq, m, ifp) \
|
|
if_handoff((struct ifqueue *)ifq, m, ifp, 0)
|
|
#define IF_HANDOFF_ADJ(ifq, m, ifp, adj) \
|
|
if_handoff((struct ifqueue *)ifq, m, ifp, adj)
|
|
|
|
void if_start(struct ifnet *);
|
|
|
|
#define IFQ_ENQUEUE(ifq, m, err) \
|
|
do { \
|
|
IF_LOCK(ifq); \
|
|
if (ALTQ_IS_ENABLED(ifq)) \
|
|
ALTQ_ENQUEUE(ifq, m, NULL, err); \
|
|
else { \
|
|
if (_IF_QFULL(ifq)) { \
|
|
m_freem(m); \
|
|
(err) = ENOBUFS; \
|
|
} else { \
|
|
_IF_ENQUEUE(ifq, m); \
|
|
(err) = 0; \
|
|
} \
|
|
} \
|
|
IF_UNLOCK(ifq); \
|
|
} while (0)
|
|
|
|
#define IFQ_DEQUEUE_NOLOCK(ifq, m) \
|
|
do { \
|
|
if (TBR_IS_ENABLED(ifq)) \
|
|
(m) = tbr_dequeue_ptr(ifq, ALTDQ_REMOVE); \
|
|
else if (ALTQ_IS_ENABLED(ifq)) \
|
|
ALTQ_DEQUEUE(ifq, m); \
|
|
else \
|
|
_IF_DEQUEUE(ifq, m); \
|
|
} while (0)
|
|
|
|
#define IFQ_DEQUEUE(ifq, m) \
|
|
do { \
|
|
IF_LOCK(ifq); \
|
|
IFQ_DEQUEUE_NOLOCK(ifq, m); \
|
|
IF_UNLOCK(ifq); \
|
|
} while (0)
|
|
|
|
#define IFQ_POLL_NOLOCK(ifq, m) \
|
|
do { \
|
|
if (TBR_IS_ENABLED(ifq)) \
|
|
(m) = tbr_dequeue_ptr(ifq, ALTDQ_POLL); \
|
|
else if (ALTQ_IS_ENABLED(ifq)) \
|
|
ALTQ_POLL(ifq, m); \
|
|
else \
|
|
_IF_POLL(ifq, m); \
|
|
} while (0)
|
|
|
|
#define IFQ_POLL(ifq, m) \
|
|
do { \
|
|
IF_LOCK(ifq); \
|
|
IFQ_POLL_NOLOCK(ifq, m); \
|
|
IF_UNLOCK(ifq); \
|
|
} while (0)
|
|
|
|
#define IFQ_PURGE_NOLOCK(ifq) \
|
|
do { \
|
|
if (ALTQ_IS_ENABLED(ifq)) { \
|
|
ALTQ_PURGE(ifq); \
|
|
} else \
|
|
_IF_DRAIN(ifq); \
|
|
} while (0)
|
|
|
|
#define IFQ_PURGE(ifq) \
|
|
do { \
|
|
IF_LOCK(ifq); \
|
|
IFQ_PURGE_NOLOCK(ifq); \
|
|
IF_UNLOCK(ifq); \
|
|
} while (0)
|
|
|
|
#define IFQ_SET_READY(ifq) \
|
|
do { ((ifq)->altq_flags |= ALTQF_READY); } while (0)
|
|
|
|
#define IFQ_LOCK(ifq) IF_LOCK(ifq)
|
|
#define IFQ_UNLOCK(ifq) IF_UNLOCK(ifq)
|
|
#define IFQ_LOCK_ASSERT(ifq) IF_LOCK_ASSERT(ifq)
|
|
#define IFQ_IS_EMPTY(ifq) ((ifq)->ifq_len == 0)
|
|
#define IFQ_INC_LEN(ifq) ((ifq)->ifq_len++)
|
|
#define IFQ_DEC_LEN(ifq) (--(ifq)->ifq_len)
|
|
#define IFQ_SET_MAXLEN(ifq, len) ((ifq)->ifq_maxlen = (len))
|
|
|
|
/*
|
|
* The IFF_DRV_OACTIVE test should really occur in the device driver, not in
|
|
* the handoff logic, as that flag is locked by the device driver.
|
|
*/
|
|
#define IFQ_HANDOFF_ADJ(ifp, m, adj, err) \
|
|
do { \
|
|
int len; \
|
|
short mflags; \
|
|
\
|
|
len = (m)->m_pkthdr.len; \
|
|
mflags = (m)->m_flags; \
|
|
IFQ_ENQUEUE(&(ifp)->if_snd, m, err); \
|
|
if ((err) == 0) { \
|
|
if_inc_counter((ifp), IFCOUNTER_OBYTES, len + (adj)); \
|
|
if (mflags & M_MCAST) \
|
|
if_inc_counter((ifp), IFCOUNTER_OMCASTS, 1); \
|
|
if (((ifp)->if_drv_flags & IFF_DRV_OACTIVE) == 0) \
|
|
if_start(ifp); \
|
|
} else \
|
|
if_inc_counter((ifp), IFCOUNTER_OQDROPS, 1); \
|
|
} while (0)
|
|
|
|
#define IFQ_HANDOFF(ifp, m, err) \
|
|
IFQ_HANDOFF_ADJ(ifp, m, 0, err)
|
|
|
|
#define IFQ_DRV_DEQUEUE(ifq, m) \
|
|
do { \
|
|
(m) = (ifq)->ifq_drv_head; \
|
|
if (m) { \
|
|
if (((ifq)->ifq_drv_head = (m)->m_nextpkt) == NULL) \
|
|
(ifq)->ifq_drv_tail = NULL; \
|
|
(m)->m_nextpkt = NULL; \
|
|
(ifq)->ifq_drv_len--; \
|
|
} else { \
|
|
IFQ_LOCK(ifq); \
|
|
IFQ_DEQUEUE_NOLOCK(ifq, m); \
|
|
while ((ifq)->ifq_drv_len < (ifq)->ifq_drv_maxlen) { \
|
|
struct mbuf *m0; \
|
|
IFQ_DEQUEUE_NOLOCK(ifq, m0); \
|
|
if (m0 == NULL) \
|
|
break; \
|
|
m0->m_nextpkt = NULL; \
|
|
if ((ifq)->ifq_drv_tail == NULL) \
|
|
(ifq)->ifq_drv_head = m0; \
|
|
else \
|
|
(ifq)->ifq_drv_tail->m_nextpkt = m0; \
|
|
(ifq)->ifq_drv_tail = m0; \
|
|
(ifq)->ifq_drv_len++; \
|
|
} \
|
|
IFQ_UNLOCK(ifq); \
|
|
} \
|
|
} while (0)
|
|
|
|
#define IFQ_DRV_PREPEND(ifq, m) \
|
|
do { \
|
|
(m)->m_nextpkt = (ifq)->ifq_drv_head; \
|
|
if ((ifq)->ifq_drv_tail == NULL) \
|
|
(ifq)->ifq_drv_tail = (m); \
|
|
(ifq)->ifq_drv_head = (m); \
|
|
(ifq)->ifq_drv_len++; \
|
|
} while (0)
|
|
|
|
#define IFQ_DRV_IS_EMPTY(ifq) \
|
|
(((ifq)->ifq_drv_len == 0) && ((ifq)->ifq_len == 0))
|
|
|
|
#define IFQ_DRV_PURGE(ifq) \
|
|
do { \
|
|
struct mbuf *m, *n = (ifq)->ifq_drv_head; \
|
|
while((m = n) != NULL) { \
|
|
n = m->m_nextpkt; \
|
|
m_freem(m); \
|
|
} \
|
|
(ifq)->ifq_drv_head = (ifq)->ifq_drv_tail = NULL; \
|
|
(ifq)->ifq_drv_len = 0; \
|
|
IFQ_PURGE(ifq); \
|
|
} while (0)
|
|
|
|
#ifdef ALTQ
|
|
int drbr_enqueue(struct ifnet *ifp, struct buf_ring *br, struct mbuf *m);
|
|
void drbr_putback(struct ifnet *ifp, struct buf_ring *br, struct mbuf *m_new);
|
|
struct mbuf *drbr_peek(struct ifnet *ifp, struct buf_ring *br);
|
|
void drbr_flush(struct ifnet *ifp, struct buf_ring *br);
|
|
struct mbuf *drbr_dequeue(struct ifnet *ifp, struct buf_ring *br);
|
|
void drbr_advance(struct ifnet *ifp, struct buf_ring *br);
|
|
struct mbuf *drbr_dequeue_cond(struct ifnet *ifp, struct buf_ring *br,
|
|
int (*func) (struct mbuf *, void *), void *arg);
|
|
int drbr_empty(struct ifnet *ifp, struct buf_ring *br);
|
|
int drbr_needs_enqueue(struct ifnet *ifp, struct buf_ring *br);
|
|
int drbr_inuse(struct ifnet *ifp, struct buf_ring *br);
|
|
|
|
#else /* !ALTQ */
|
|
static __inline int
|
|
drbr_enqueue(struct ifnet *ifp, struct buf_ring *br, struct mbuf *m)
|
|
{
|
|
int error = 0;
|
|
|
|
error = buf_ring_enqueue(br, m);
|
|
if (error)
|
|
m_freem(m);
|
|
|
|
return (error);
|
|
}
|
|
|
|
static __inline void
|
|
drbr_putback(struct ifnet *ifp, struct buf_ring *br, struct mbuf *m_new)
|
|
{
|
|
/*
|
|
* The top of the list needs to be swapped
|
|
* for this one.
|
|
*/
|
|
buf_ring_putback_sc(br, m_new);
|
|
}
|
|
|
|
static __inline struct mbuf *
|
|
drbr_peek(struct ifnet *ifp, struct buf_ring *br)
|
|
{
|
|
return ((struct mbuf *)buf_ring_peek_clear_sc(br));
|
|
}
|
|
|
|
static __inline void
|
|
drbr_flush(struct ifnet *ifp, struct buf_ring *br)
|
|
{
|
|
struct mbuf *m;
|
|
|
|
while ((m = (struct mbuf *)buf_ring_dequeue_sc(br)) != NULL)
|
|
m_freem(m);
|
|
}
|
|
|
|
static __inline struct mbuf *
|
|
drbr_dequeue(struct ifnet *ifp, struct buf_ring *br)
|
|
{
|
|
return ((struct mbuf *)buf_ring_dequeue_sc(br));
|
|
}
|
|
|
|
static __inline void
|
|
drbr_advance(struct ifnet *ifp, struct buf_ring *br)
|
|
{
|
|
return (buf_ring_advance_sc(br));
|
|
}
|
|
|
|
static __inline struct mbuf *
|
|
drbr_dequeue_cond(struct ifnet *ifp, struct buf_ring *br,
|
|
int (*func) (struct mbuf *, void *), void *arg)
|
|
{
|
|
struct mbuf *m;
|
|
m = (struct mbuf *)buf_ring_peek(br);
|
|
if (m == NULL || func(m, arg) == 0)
|
|
return (NULL);
|
|
|
|
return ((struct mbuf *)buf_ring_dequeue_sc(br));
|
|
}
|
|
|
|
static __inline int
|
|
drbr_empty(struct ifnet *ifp, struct buf_ring *br)
|
|
{
|
|
return (buf_ring_empty(br));
|
|
}
|
|
|
|
static __inline int
|
|
drbr_needs_enqueue(struct ifnet *ifp, struct buf_ring *br)
|
|
{
|
|
return (!buf_ring_empty(br));
|
|
}
|
|
|
|
static __inline int
|
|
drbr_inuse(struct ifnet *ifp, struct buf_ring *br)
|
|
{
|
|
return (buf_ring_count(br));
|
|
}
|
|
#endif /* ALTQ */
|
|
|
|
static __inline void
|
|
drbr_free(struct buf_ring *br, struct malloc_type *type)
|
|
{
|
|
|
|
drbr_flush(NULL, br);
|
|
buf_ring_free(br, type);
|
|
}
|
|
|
|
extern int ifqmaxlen;
|
|
|
|
void if_qflush(struct ifnet *);
|
|
void ifq_init(struct ifaltq *, struct ifnet *ifp);
|
|
void ifq_delete(struct ifaltq *);
|
|
|
|
#endif /* _KERNEL */
|
|
#endif /* !_NET_IFQ_H_ */
|