mirror of
https://github.com/torvalds/linux.git
synced 2024-11-10 22:21:40 +00:00
Merge git://git.kernel.org/pub/scm/linux/kernel/git/davem/net-2.6
* git://git.kernel.org/pub/scm/linux/kernel/git/davem/net-2.6: netns: fix ip_rt_frag_needed rt_is_expired netfilter: nf_conntrack_extend: avoid unnecessary "ct->ext" dereferences netfilter: fix double-free and use-after free netfilter: arptables in netns for real netfilter: ip{,6}tables_security: fix future section mismatch selinux: use nf_register_hooks() netfilter: ebtables: use nf_register_hooks() Revert "pkt_sched: sch_sfq: dump a real number of flows" qeth: use dev->ml_priv instead of dev->priv syncookies: Make sure ECN is disabled net: drop unused BUG_TRAP() net: convert BUG_TRAP to generic WARN_ON drivers/net: convert BUG_TRAP to generic WARN_ON
This commit is contained in:
commit
2284284281
@ -814,7 +814,7 @@ static u16 bnx2x_free_tx_pkt(struct bnx2x *bp, struct bnx2x_fastpath *fp,
|
||||
}
|
||||
|
||||
/* release skb */
|
||||
BUG_TRAP(skb);
|
||||
WARN_ON(!skb);
|
||||
dev_kfree_skb(skb);
|
||||
tx_buf->first_bd = 0;
|
||||
tx_buf->skb = NULL;
|
||||
@ -837,9 +837,9 @@ static inline u16 bnx2x_tx_avail(struct bnx2x_fastpath *fp)
|
||||
used = SUB_S16(prod, cons) + (s16)NUM_TX_RINGS;
|
||||
|
||||
#ifdef BNX2X_STOP_ON_ERROR
|
||||
BUG_TRAP(used >= 0);
|
||||
BUG_TRAP(used <= fp->bp->tx_ring_size);
|
||||
BUG_TRAP((fp->bp->tx_ring_size - used) <= MAX_TX_AVAIL);
|
||||
WARN_ON(used < 0);
|
||||
WARN_ON(used > fp->bp->tx_ring_size);
|
||||
WARN_ON((fp->bp->tx_ring_size - used) > MAX_TX_AVAIL);
|
||||
#endif
|
||||
|
||||
return (s16)(fp->bp->tx_ring_size) - used;
|
||||
@ -4374,7 +4374,7 @@ static void bnx2x_init_rx_rings(struct bnx2x *bp)
|
||||
}
|
||||
ring_prod = NEXT_RX_IDX(ring_prod);
|
||||
cqe_ring_prod = NEXT_RCQ_IDX(cqe_ring_prod);
|
||||
BUG_TRAP(ring_prod > i);
|
||||
WARN_ON(ring_prod <= i);
|
||||
}
|
||||
|
||||
fp->rx_bd_prod = ring_prod;
|
||||
|
@ -3321,7 +3321,7 @@ int qeth_change_mtu(struct net_device *dev, int new_mtu)
|
||||
struct qeth_card *card;
|
||||
char dbf_text[15];
|
||||
|
||||
card = netdev_priv(dev);
|
||||
card = dev->ml_priv;
|
||||
|
||||
QETH_DBF_TEXT(TRACE, 4, "chgmtu");
|
||||
sprintf(dbf_text, "%8x", new_mtu);
|
||||
@ -3343,7 +3343,7 @@ struct net_device_stats *qeth_get_stats(struct net_device *dev)
|
||||
{
|
||||
struct qeth_card *card;
|
||||
|
||||
card = netdev_priv(dev);
|
||||
card = dev->ml_priv;
|
||||
|
||||
QETH_DBF_TEXT(TRACE, 5, "getstat");
|
||||
|
||||
@ -3395,7 +3395,7 @@ void qeth_tx_timeout(struct net_device *dev)
|
||||
{
|
||||
struct qeth_card *card;
|
||||
|
||||
card = netdev_priv(dev);
|
||||
card = dev->ml_priv;
|
||||
card->stats.tx_errors++;
|
||||
qeth_schedule_recovery(card);
|
||||
}
|
||||
@ -3403,7 +3403,7 @@ EXPORT_SYMBOL_GPL(qeth_tx_timeout);
|
||||
|
||||
int qeth_mdio_read(struct net_device *dev, int phy_id, int regnum)
|
||||
{
|
||||
struct qeth_card *card = netdev_priv(dev);
|
||||
struct qeth_card *card = dev->ml_priv;
|
||||
int rc = 0;
|
||||
|
||||
switch (regnum) {
|
||||
@ -4253,7 +4253,7 @@ EXPORT_SYMBOL_GPL(qeth_core_get_stats_count);
|
||||
void qeth_core_get_ethtool_stats(struct net_device *dev,
|
||||
struct ethtool_stats *stats, u64 *data)
|
||||
{
|
||||
struct qeth_card *card = netdev_priv(dev);
|
||||
struct qeth_card *card = dev->ml_priv;
|
||||
data[0] = card->stats.rx_packets -
|
||||
card->perf_stats.initial_rx_packets;
|
||||
data[1] = card->perf_stats.bufs_rec;
|
||||
@ -4313,7 +4313,7 @@ EXPORT_SYMBOL_GPL(qeth_core_get_strings);
|
||||
void qeth_core_get_drvinfo(struct net_device *dev,
|
||||
struct ethtool_drvinfo *info)
|
||||
{
|
||||
struct qeth_card *card = netdev_priv(dev);
|
||||
struct qeth_card *card = dev->ml_priv;
|
||||
if (card->options.layer2)
|
||||
strcpy(info->driver, "qeth_l2");
|
||||
else
|
||||
@ -4331,7 +4331,7 @@ EXPORT_SYMBOL_GPL(qeth_core_get_drvinfo);
|
||||
int qeth_core_ethtool_get_settings(struct net_device *netdev,
|
||||
struct ethtool_cmd *ecmd)
|
||||
{
|
||||
struct qeth_card *card = netdev_priv(netdev);
|
||||
struct qeth_card *card = netdev->ml_priv;
|
||||
enum qeth_link_types link_type;
|
||||
|
||||
if ((card->info.type == QETH_CARD_TYPE_IQD) || (card->info.guestlan))
|
||||
|
@ -35,7 +35,7 @@ static int qeth_l2_recover(void *);
|
||||
|
||||
static int qeth_l2_do_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
|
||||
{
|
||||
struct qeth_card *card = netdev_priv(dev);
|
||||
struct qeth_card *card = dev->ml_priv;
|
||||
struct mii_ioctl_data *mii_data;
|
||||
int rc = 0;
|
||||
|
||||
@ -317,7 +317,7 @@ static void qeth_l2_process_vlans(struct qeth_card *card, int clear)
|
||||
|
||||
static void qeth_l2_vlan_rx_add_vid(struct net_device *dev, unsigned short vid)
|
||||
{
|
||||
struct qeth_card *card = netdev_priv(dev);
|
||||
struct qeth_card *card = dev->ml_priv;
|
||||
struct qeth_vlan_vid *id;
|
||||
|
||||
QETH_DBF_TEXT_(TRACE, 4, "aid:%d", vid);
|
||||
@ -334,7 +334,7 @@ static void qeth_l2_vlan_rx_add_vid(struct net_device *dev, unsigned short vid)
|
||||
static void qeth_l2_vlan_rx_kill_vid(struct net_device *dev, unsigned short vid)
|
||||
{
|
||||
struct qeth_vlan_vid *id, *tmpid = NULL;
|
||||
struct qeth_card *card = netdev_priv(dev);
|
||||
struct qeth_card *card = dev->ml_priv;
|
||||
|
||||
QETH_DBF_TEXT_(TRACE, 4, "kid:%d", vid);
|
||||
spin_lock_bh(&card->vlanlock);
|
||||
@ -566,7 +566,7 @@ static int qeth_l2_request_initial_mac(struct qeth_card *card)
|
||||
static int qeth_l2_set_mac_address(struct net_device *dev, void *p)
|
||||
{
|
||||
struct sockaddr *addr = p;
|
||||
struct qeth_card *card = netdev_priv(dev);
|
||||
struct qeth_card *card = dev->ml_priv;
|
||||
int rc = 0;
|
||||
|
||||
QETH_DBF_TEXT(TRACE, 3, "setmac");
|
||||
@ -590,7 +590,7 @@ static int qeth_l2_set_mac_address(struct net_device *dev, void *p)
|
||||
|
||||
static void qeth_l2_set_multicast_list(struct net_device *dev)
|
||||
{
|
||||
struct qeth_card *card = netdev_priv(dev);
|
||||
struct qeth_card *card = dev->ml_priv;
|
||||
struct dev_mc_list *dm;
|
||||
|
||||
if (card->info.type == QETH_CARD_TYPE_OSN)
|
||||
@ -612,7 +612,7 @@ static int qeth_l2_hard_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||
int rc;
|
||||
struct qeth_hdr *hdr = NULL;
|
||||
int elements = 0;
|
||||
struct qeth_card *card = netdev_priv(dev);
|
||||
struct qeth_card *card = dev->ml_priv;
|
||||
struct sk_buff *new_skb = skb;
|
||||
int ipv = qeth_get_ip_version(skb);
|
||||
int cast_type = qeth_get_cast_type(card, skb);
|
||||
@ -767,7 +767,7 @@ static void qeth_l2_qdio_input_handler(struct ccw_device *ccwdev,
|
||||
|
||||
static int qeth_l2_open(struct net_device *dev)
|
||||
{
|
||||
struct qeth_card *card = netdev_priv(dev);
|
||||
struct qeth_card *card = dev->ml_priv;
|
||||
|
||||
QETH_DBF_TEXT(TRACE, 4, "qethopen");
|
||||
if (card->state != CARD_STATE_SOFTSETUP)
|
||||
@ -791,7 +791,7 @@ static int qeth_l2_open(struct net_device *dev)
|
||||
|
||||
static int qeth_l2_stop(struct net_device *dev)
|
||||
{
|
||||
struct qeth_card *card = netdev_priv(dev);
|
||||
struct qeth_card *card = dev->ml_priv;
|
||||
|
||||
QETH_DBF_TEXT(TRACE, 4, "qethstop");
|
||||
netif_tx_disable(dev);
|
||||
@ -838,7 +838,7 @@ static void qeth_l2_remove_device(struct ccwgroup_device *cgdev)
|
||||
|
||||
static int qeth_l2_ethtool_set_tso(struct net_device *dev, u32 data)
|
||||
{
|
||||
struct qeth_card *card = netdev_priv(dev);
|
||||
struct qeth_card *card = dev->ml_priv;
|
||||
|
||||
if (data) {
|
||||
if (card->options.large_send == QETH_LARGE_SEND_NO) {
|
||||
@ -894,7 +894,7 @@ static int qeth_l2_setup_netdev(struct qeth_card *card)
|
||||
if (!card->dev)
|
||||
return -ENODEV;
|
||||
|
||||
card->dev->priv = card;
|
||||
card->dev->ml_priv = card;
|
||||
card->dev->tx_timeout = &qeth_tx_timeout;
|
||||
card->dev->watchdog_timeo = QETH_TX_TIMEOUT;
|
||||
card->dev->open = qeth_l2_open;
|
||||
@ -1178,7 +1178,7 @@ int qeth_osn_assist(struct net_device *dev, void *data, int data_len)
|
||||
QETH_DBF_TEXT(TRACE, 2, "osnsdmc");
|
||||
if (!dev)
|
||||
return -ENODEV;
|
||||
card = netdev_priv(dev);
|
||||
card = dev->ml_priv;
|
||||
if (!card)
|
||||
return -ENODEV;
|
||||
if ((card->state != CARD_STATE_UP) &&
|
||||
@ -1201,7 +1201,7 @@ int qeth_osn_register(unsigned char *read_dev_no, struct net_device **dev,
|
||||
*dev = qeth_l2_netdev_by_devno(read_dev_no);
|
||||
if (*dev == NULL)
|
||||
return -ENODEV;
|
||||
card = netdev_priv(*dev);
|
||||
card = (*dev)->ml_priv;
|
||||
if (!card)
|
||||
return -ENODEV;
|
||||
if ((assist_cb == NULL) || (data_cb == NULL))
|
||||
@ -1219,7 +1219,7 @@ void qeth_osn_deregister(struct net_device *dev)
|
||||
QETH_DBF_TEXT(TRACE, 2, "osndereg");
|
||||
if (!dev)
|
||||
return;
|
||||
card = netdev_priv(dev);
|
||||
card = dev->ml_priv;
|
||||
if (!card)
|
||||
return;
|
||||
card->osn_info.assist_cb = NULL;
|
||||
|
@ -1813,7 +1813,7 @@ static void qeth_l3_free_vlan_addresses(struct qeth_card *card,
|
||||
static void qeth_l3_vlan_rx_register(struct net_device *dev,
|
||||
struct vlan_group *grp)
|
||||
{
|
||||
struct qeth_card *card = netdev_priv(dev);
|
||||
struct qeth_card *card = dev->ml_priv;
|
||||
unsigned long flags;
|
||||
|
||||
QETH_DBF_TEXT(TRACE, 4, "vlanreg");
|
||||
@ -1825,7 +1825,7 @@ static void qeth_l3_vlan_rx_register(struct net_device *dev,
|
||||
static void qeth_l3_vlan_rx_add_vid(struct net_device *dev, unsigned short vid)
|
||||
{
|
||||
struct net_device *vlandev;
|
||||
struct qeth_card *card = (struct qeth_card *) dev->priv;
|
||||
struct qeth_card *card = dev->ml_priv;
|
||||
struct in_device *in_dev;
|
||||
|
||||
if (card->info.type == QETH_CARD_TYPE_IQD)
|
||||
@ -1851,7 +1851,7 @@ static void qeth_l3_vlan_rx_add_vid(struct net_device *dev, unsigned short vid)
|
||||
|
||||
static void qeth_l3_vlan_rx_kill_vid(struct net_device *dev, unsigned short vid)
|
||||
{
|
||||
struct qeth_card *card = netdev_priv(dev);
|
||||
struct qeth_card *card = dev->ml_priv;
|
||||
unsigned long flags;
|
||||
|
||||
QETH_DBF_TEXT_(TRACE, 4, "kid:%d", vid);
|
||||
@ -2013,7 +2013,7 @@ static int qeth_l3_verify_vlan_dev(struct net_device *dev,
|
||||
}
|
||||
}
|
||||
|
||||
if (rc && !(netdev_priv(vlan_dev_real_dev(dev)) == (void *)card))
|
||||
if (rc && !(vlan_dev_real_dev(dev)->ml_priv == (void *)card))
|
||||
return 0;
|
||||
|
||||
return rc;
|
||||
@ -2047,9 +2047,9 @@ static struct qeth_card *qeth_l3_get_card_from_dev(struct net_device *dev)
|
||||
|
||||
rc = qeth_l3_verify_dev(dev);
|
||||
if (rc == QETH_REAL_CARD)
|
||||
card = netdev_priv(dev);
|
||||
card = dev->ml_priv;
|
||||
else if (rc == QETH_VLAN_CARD)
|
||||
card = netdev_priv(vlan_dev_real_dev(dev));
|
||||
card = vlan_dev_real_dev(dev)->ml_priv;
|
||||
if (card && card->options.layer2)
|
||||
card = NULL;
|
||||
QETH_DBF_TEXT_(TRACE, 4, "%d", rc);
|
||||
@ -2110,7 +2110,7 @@ static int qeth_l3_stop_card(struct qeth_card *card, int recovery_mode)
|
||||
|
||||
static void qeth_l3_set_multicast_list(struct net_device *dev)
|
||||
{
|
||||
struct qeth_card *card = netdev_priv(dev);
|
||||
struct qeth_card *card = dev->ml_priv;
|
||||
|
||||
QETH_DBF_TEXT(TRACE, 3, "setmulti");
|
||||
qeth_l3_delete_mc_addresses(card);
|
||||
@ -2438,7 +2438,7 @@ static int qeth_l3_arp_flush_cache(struct qeth_card *card)
|
||||
|
||||
static int qeth_l3_do_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
|
||||
{
|
||||
struct qeth_card *card = netdev_priv(dev);
|
||||
struct qeth_card *card = dev->ml_priv;
|
||||
struct qeth_arp_cache_entry arp_entry;
|
||||
struct mii_ioctl_data *mii_data;
|
||||
int rc = 0;
|
||||
@ -2595,7 +2595,7 @@ static int qeth_l3_hard_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||
u16 *tag;
|
||||
struct qeth_hdr *hdr = NULL;
|
||||
int elements_needed = 0;
|
||||
struct qeth_card *card = netdev_priv(dev);
|
||||
struct qeth_card *card = dev->ml_priv;
|
||||
struct sk_buff *new_skb = NULL;
|
||||
int ipv = qeth_get_ip_version(skb);
|
||||
int cast_type = qeth_get_cast_type(card, skb);
|
||||
@ -2763,7 +2763,7 @@ tx_drop:
|
||||
|
||||
static int qeth_l3_open(struct net_device *dev)
|
||||
{
|
||||
struct qeth_card *card = netdev_priv(dev);
|
||||
struct qeth_card *card = dev->ml_priv;
|
||||
|
||||
QETH_DBF_TEXT(TRACE, 4, "qethopen");
|
||||
if (card->state != CARD_STATE_SOFTSETUP)
|
||||
@ -2780,7 +2780,7 @@ static int qeth_l3_open(struct net_device *dev)
|
||||
|
||||
static int qeth_l3_stop(struct net_device *dev)
|
||||
{
|
||||
struct qeth_card *card = netdev_priv(dev);
|
||||
struct qeth_card *card = dev->ml_priv;
|
||||
|
||||
QETH_DBF_TEXT(TRACE, 4, "qethstop");
|
||||
netif_tx_disable(dev);
|
||||
@ -2792,14 +2792,14 @@ static int qeth_l3_stop(struct net_device *dev)
|
||||
|
||||
static u32 qeth_l3_ethtool_get_rx_csum(struct net_device *dev)
|
||||
{
|
||||
struct qeth_card *card = netdev_priv(dev);
|
||||
struct qeth_card *card = dev->ml_priv;
|
||||
|
||||
return (card->options.checksum_type == HW_CHECKSUMMING);
|
||||
}
|
||||
|
||||
static int qeth_l3_ethtool_set_rx_csum(struct net_device *dev, u32 data)
|
||||
{
|
||||
struct qeth_card *card = netdev_priv(dev);
|
||||
struct qeth_card *card = dev->ml_priv;
|
||||
enum qeth_card_states old_state;
|
||||
enum qeth_checksum_types csum_type;
|
||||
|
||||
@ -2825,7 +2825,7 @@ static int qeth_l3_ethtool_set_rx_csum(struct net_device *dev, u32 data)
|
||||
|
||||
static int qeth_l3_ethtool_set_tso(struct net_device *dev, u32 data)
|
||||
{
|
||||
struct qeth_card *card = netdev_priv(dev);
|
||||
struct qeth_card *card = dev->ml_priv;
|
||||
|
||||
if (data) {
|
||||
if (card->options.large_send == QETH_LARGE_SEND_NO) {
|
||||
@ -2915,7 +2915,7 @@ static int qeth_l3_setup_netdev(struct qeth_card *card)
|
||||
return -ENODEV;
|
||||
|
||||
card->dev->hard_start_xmit = qeth_l3_hard_start_xmit;
|
||||
card->dev->priv = card;
|
||||
card->dev->ml_priv = card;
|
||||
card->dev->tx_timeout = &qeth_tx_timeout;
|
||||
card->dev->watchdog_timeo = QETH_TX_TIMEOUT;
|
||||
card->dev->open = qeth_l3_open;
|
||||
|
@ -755,13 +755,6 @@ extern void __rtnl_unlock(void);
|
||||
} \
|
||||
} while(0)
|
||||
|
||||
#define BUG_TRAP(x) do { \
|
||||
if (unlikely(!(x))) { \
|
||||
printk(KERN_ERR "KERNEL: assertion (%s) failed at %s (%d)\n", \
|
||||
#x, __FILE__ , __LINE__); \
|
||||
} \
|
||||
} while(0)
|
||||
|
||||
static inline u32 rtm_get_table(struct rtattr **rta, u8 table)
|
||||
{
|
||||
return RTA_GET_U32(rta[RTA_TABLE-1]);
|
||||
|
@ -96,6 +96,7 @@ int kmem_ptr_validate(struct kmem_cache *cachep, const void *ptr);
|
||||
/*
|
||||
* Common kmalloc functions provided by all allocators
|
||||
*/
|
||||
void * __must_check __krealloc(const void *, size_t, gfp_t);
|
||||
void * __must_check krealloc(const void *, size_t, gfp_t);
|
||||
void kfree(const void *);
|
||||
size_t ksize(const void *);
|
||||
|
@ -18,6 +18,7 @@
|
||||
#include <linux/slab.h>
|
||||
#include <linux/spinlock.h>
|
||||
#include <linux/types.h>
|
||||
#include <linux/bug.h>
|
||||
|
||||
#include <net/sock.h>
|
||||
|
||||
@ -170,7 +171,7 @@ static inline struct request_sock *reqsk_queue_remove(struct request_sock_queue
|
||||
{
|
||||
struct request_sock *req = queue->rskq_accept_head;
|
||||
|
||||
BUG_TRAP(req != NULL);
|
||||
WARN_ON(req == NULL);
|
||||
|
||||
queue->rskq_accept_head = req->dl_next;
|
||||
if (queue->rskq_accept_head == NULL)
|
||||
@ -185,7 +186,7 @@ static inline struct sock *reqsk_queue_get_child(struct request_sock_queue *queu
|
||||
struct request_sock *req = reqsk_queue_remove(queue);
|
||||
struct sock *child = req->sk;
|
||||
|
||||
BUG_TRAP(child != NULL);
|
||||
WARN_ON(child == NULL);
|
||||
|
||||
sk_acceptq_removed(parent);
|
||||
__reqsk_free(req);
|
||||
|
46
mm/util.c
46
mm/util.c
@ -69,6 +69,38 @@ void *kmemdup(const void *src, size_t len, gfp_t gfp)
|
||||
}
|
||||
EXPORT_SYMBOL(kmemdup);
|
||||
|
||||
/**
|
||||
* __krealloc - like krealloc() but don't free @p.
|
||||
* @p: object to reallocate memory for.
|
||||
* @new_size: how many bytes of memory are required.
|
||||
* @flags: the type of memory to allocate.
|
||||
*
|
||||
* This function is like krealloc() except it never frees the originally
|
||||
* allocated buffer. Use this if you don't want to free the buffer immediately
|
||||
* like, for example, with RCU.
|
||||
*/
|
||||
void *__krealloc(const void *p, size_t new_size, gfp_t flags)
|
||||
{
|
||||
void *ret;
|
||||
size_t ks = 0;
|
||||
|
||||
if (unlikely(!new_size))
|
||||
return ZERO_SIZE_PTR;
|
||||
|
||||
if (p)
|
||||
ks = ksize(p);
|
||||
|
||||
if (ks >= new_size)
|
||||
return (void *)p;
|
||||
|
||||
ret = kmalloc_track_caller(new_size, flags);
|
||||
if (ret && p)
|
||||
memcpy(ret, p, ks);
|
||||
|
||||
return ret;
|
||||
}
|
||||
EXPORT_SYMBOL(__krealloc);
|
||||
|
||||
/**
|
||||
* krealloc - reallocate memory. The contents will remain unchanged.
|
||||
* @p: object to reallocate memory for.
|
||||
@ -83,24 +115,16 @@ EXPORT_SYMBOL(kmemdup);
|
||||
void *krealloc(const void *p, size_t new_size, gfp_t flags)
|
||||
{
|
||||
void *ret;
|
||||
size_t ks = 0;
|
||||
|
||||
if (unlikely(!new_size)) {
|
||||
kfree(p);
|
||||
return ZERO_SIZE_PTR;
|
||||
}
|
||||
|
||||
if (p)
|
||||
ks = ksize(p);
|
||||
|
||||
if (ks >= new_size)
|
||||
return (void *)p;
|
||||
|
||||
ret = kmalloc_track_caller(new_size, flags);
|
||||
if (ret && p) {
|
||||
memcpy(ret, p, ks);
|
||||
ret = __krealloc(p, new_size, flags);
|
||||
if (ret && p != ret)
|
||||
kfree(p);
|
||||
}
|
||||
|
||||
return ret;
|
||||
}
|
||||
EXPORT_SYMBOL(krealloc);
|
||||
|
@ -959,7 +959,7 @@ static unsigned long atalk_sum_skb(const struct sk_buff *skb, int offset,
|
||||
for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) {
|
||||
int end;
|
||||
|
||||
BUG_TRAP(start <= offset + len);
|
||||
WARN_ON(start > offset + len);
|
||||
|
||||
end = start + skb_shinfo(skb)->frags[i].size;
|
||||
if ((copy = end - offset) > 0) {
|
||||
@ -986,7 +986,7 @@ static unsigned long atalk_sum_skb(const struct sk_buff *skb, int offset,
|
||||
for (; list; list = list->next) {
|
||||
int end;
|
||||
|
||||
BUG_TRAP(start <= offset + len);
|
||||
WARN_ON(start > offset + len);
|
||||
|
||||
end = start + list->len;
|
||||
if ((copy = end - offset) > 0) {
|
||||
|
@ -93,28 +93,20 @@ static struct nf_hook_ops ebt_ops_filter[] __read_mostly = {
|
||||
|
||||
static int __init ebtable_filter_init(void)
|
||||
{
|
||||
int i, j, ret;
|
||||
int ret;
|
||||
|
||||
ret = ebt_register_table(&frame_filter);
|
||||
if (ret < 0)
|
||||
return ret;
|
||||
for (i = 0; i < ARRAY_SIZE(ebt_ops_filter); i++)
|
||||
if ((ret = nf_register_hook(&ebt_ops_filter[i])) < 0)
|
||||
goto cleanup;
|
||||
return ret;
|
||||
cleanup:
|
||||
for (j = 0; j < i; j++)
|
||||
nf_unregister_hook(&ebt_ops_filter[j]);
|
||||
ebt_unregister_table(&frame_filter);
|
||||
ret = nf_register_hooks(ebt_ops_filter, ARRAY_SIZE(ebt_ops_filter));
|
||||
if (ret < 0)
|
||||
ebt_unregister_table(&frame_filter);
|
||||
return ret;
|
||||
}
|
||||
|
||||
static void __exit ebtable_filter_fini(void)
|
||||
{
|
||||
int i;
|
||||
|
||||
for (i = 0; i < ARRAY_SIZE(ebt_ops_filter); i++)
|
||||
nf_unregister_hook(&ebt_ops_filter[i]);
|
||||
nf_unregister_hooks(ebt_ops_filter, ARRAY_SIZE(ebt_ops_filter));
|
||||
ebt_unregister_table(&frame_filter);
|
||||
}
|
||||
|
||||
|
@ -100,28 +100,20 @@ static struct nf_hook_ops ebt_ops_nat[] __read_mostly = {
|
||||
|
||||
static int __init ebtable_nat_init(void)
|
||||
{
|
||||
int i, ret, j;
|
||||
int ret;
|
||||
|
||||
ret = ebt_register_table(&frame_nat);
|
||||
if (ret < 0)
|
||||
return ret;
|
||||
for (i = 0; i < ARRAY_SIZE(ebt_ops_nat); i++)
|
||||
if ((ret = nf_register_hook(&ebt_ops_nat[i])) < 0)
|
||||
goto cleanup;
|
||||
return ret;
|
||||
cleanup:
|
||||
for (j = 0; j < i; j++)
|
||||
nf_unregister_hook(&ebt_ops_nat[j]);
|
||||
ebt_unregister_table(&frame_nat);
|
||||
ret = nf_register_hooks(ebt_ops_nat, ARRAY_SIZE(ebt_ops_nat));
|
||||
if (ret < 0)
|
||||
ebt_unregister_table(&frame_nat);
|
||||
return ret;
|
||||
}
|
||||
|
||||
static void __exit ebtable_nat_fini(void)
|
||||
{
|
||||
int i;
|
||||
|
||||
for (i = 0; i < ARRAY_SIZE(ebt_ops_nat); i++)
|
||||
nf_unregister_hook(&ebt_ops_nat[i]);
|
||||
nf_unregister_hooks(ebt_ops_nat, ARRAY_SIZE(ebt_ops_nat));
|
||||
ebt_unregister_table(&frame_nat);
|
||||
}
|
||||
|
||||
|
@ -285,7 +285,7 @@ int skb_copy_datagram_iovec(const struct sk_buff *skb, int offset,
|
||||
for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) {
|
||||
int end;
|
||||
|
||||
BUG_TRAP(start <= offset + len);
|
||||
WARN_ON(start > offset + len);
|
||||
|
||||
end = start + skb_shinfo(skb)->frags[i].size;
|
||||
if ((copy = end - offset) > 0) {
|
||||
@ -315,7 +315,7 @@ int skb_copy_datagram_iovec(const struct sk_buff *skb, int offset,
|
||||
for (; list; list = list->next) {
|
||||
int end;
|
||||
|
||||
BUG_TRAP(start <= offset + len);
|
||||
WARN_ON(start > offset + len);
|
||||
|
||||
end = start + list->len;
|
||||
if ((copy = end - offset) > 0) {
|
||||
@ -366,7 +366,7 @@ static int skb_copy_and_csum_datagram(const struct sk_buff *skb, int offset,
|
||||
for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) {
|
||||
int end;
|
||||
|
||||
BUG_TRAP(start <= offset + len);
|
||||
WARN_ON(start > offset + len);
|
||||
|
||||
end = start + skb_shinfo(skb)->frags[i].size;
|
||||
if ((copy = end - offset) > 0) {
|
||||
@ -402,7 +402,7 @@ static int skb_copy_and_csum_datagram(const struct sk_buff *skb, int offset,
|
||||
for (; list; list=list->next) {
|
||||
int end;
|
||||
|
||||
BUG_TRAP(start <= offset + len);
|
||||
WARN_ON(start > offset + len);
|
||||
|
||||
end = start + list->len;
|
||||
if ((copy = end - offset) > 0) {
|
||||
|
@ -1973,7 +1973,7 @@ static void net_tx_action(struct softirq_action *h)
|
||||
struct sk_buff *skb = clist;
|
||||
clist = clist->next;
|
||||
|
||||
BUG_TRAP(!atomic_read(&skb->users));
|
||||
WARN_ON(atomic_read(&skb->users));
|
||||
__kfree_skb(skb);
|
||||
}
|
||||
}
|
||||
@ -3847,7 +3847,7 @@ static void rollback_registered(struct net_device *dev)
|
||||
dev->uninit(dev);
|
||||
|
||||
/* Notifier chain MUST detach us from master device. */
|
||||
BUG_TRAP(!dev->master);
|
||||
WARN_ON(dev->master);
|
||||
|
||||
/* Remove entries from kobject tree */
|
||||
netdev_unregister_kobject(dev);
|
||||
@ -4169,9 +4169,9 @@ void netdev_run_todo(void)
|
||||
|
||||
/* paranoia */
|
||||
BUG_ON(atomic_read(&dev->refcnt));
|
||||
BUG_TRAP(!dev->ip_ptr);
|
||||
BUG_TRAP(!dev->ip6_ptr);
|
||||
BUG_TRAP(!dev->dn_ptr);
|
||||
WARN_ON(dev->ip_ptr);
|
||||
WARN_ON(dev->ip6_ptr);
|
||||
WARN_ON(dev->dn_ptr);
|
||||
|
||||
if (dev->destructor)
|
||||
dev->destructor(dev);
|
||||
|
@ -123,7 +123,7 @@ void reqsk_queue_destroy(struct request_sock_queue *queue)
|
||||
}
|
||||
}
|
||||
|
||||
BUG_TRAP(lopt->qlen == 0);
|
||||
WARN_ON(lopt->qlen != 0);
|
||||
if (lopt_size > PAGE_SIZE)
|
||||
vfree(lopt);
|
||||
else
|
||||
|
@ -1200,7 +1200,7 @@ int skb_copy_bits(const struct sk_buff *skb, int offset, void *to, int len)
|
||||
for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) {
|
||||
int end;
|
||||
|
||||
BUG_TRAP(start <= offset + len);
|
||||
WARN_ON(start > offset + len);
|
||||
|
||||
end = start + skb_shinfo(skb)->frags[i].size;
|
||||
if ((copy = end - offset) > 0) {
|
||||
@ -1229,7 +1229,7 @@ int skb_copy_bits(const struct sk_buff *skb, int offset, void *to, int len)
|
||||
for (; list; list = list->next) {
|
||||
int end;
|
||||
|
||||
BUG_TRAP(start <= offset + len);
|
||||
WARN_ON(start > offset + len);
|
||||
|
||||
end = start + list->len;
|
||||
if ((copy = end - offset) > 0) {
|
||||
@ -1475,7 +1475,7 @@ int skb_store_bits(struct sk_buff *skb, int offset, const void *from, int len)
|
||||
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
|
||||
int end;
|
||||
|
||||
BUG_TRAP(start <= offset + len);
|
||||
WARN_ON(start > offset + len);
|
||||
|
||||
end = start + frag->size;
|
||||
if ((copy = end - offset) > 0) {
|
||||
@ -1503,7 +1503,7 @@ int skb_store_bits(struct sk_buff *skb, int offset, const void *from, int len)
|
||||
for (; list; list = list->next) {
|
||||
int end;
|
||||
|
||||
BUG_TRAP(start <= offset + len);
|
||||
WARN_ON(start > offset + len);
|
||||
|
||||
end = start + list->len;
|
||||
if ((copy = end - offset) > 0) {
|
||||
@ -1552,7 +1552,7 @@ __wsum skb_checksum(const struct sk_buff *skb, int offset,
|
||||
for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) {
|
||||
int end;
|
||||
|
||||
BUG_TRAP(start <= offset + len);
|
||||
WARN_ON(start > offset + len);
|
||||
|
||||
end = start + skb_shinfo(skb)->frags[i].size;
|
||||
if ((copy = end - offset) > 0) {
|
||||
@ -1581,7 +1581,7 @@ __wsum skb_checksum(const struct sk_buff *skb, int offset,
|
||||
for (; list; list = list->next) {
|
||||
int end;
|
||||
|
||||
BUG_TRAP(start <= offset + len);
|
||||
WARN_ON(start > offset + len);
|
||||
|
||||
end = start + list->len;
|
||||
if ((copy = end - offset) > 0) {
|
||||
@ -1629,7 +1629,7 @@ __wsum skb_copy_and_csum_bits(const struct sk_buff *skb, int offset,
|
||||
for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) {
|
||||
int end;
|
||||
|
||||
BUG_TRAP(start <= offset + len);
|
||||
WARN_ON(start > offset + len);
|
||||
|
||||
end = start + skb_shinfo(skb)->frags[i].size;
|
||||
if ((copy = end - offset) > 0) {
|
||||
@ -1662,7 +1662,7 @@ __wsum skb_copy_and_csum_bits(const struct sk_buff *skb, int offset,
|
||||
__wsum csum2;
|
||||
int end;
|
||||
|
||||
BUG_TRAP(start <= offset + len);
|
||||
WARN_ON(start > offset + len);
|
||||
|
||||
end = start + list->len;
|
||||
if ((copy = end - offset) > 0) {
|
||||
@ -2373,7 +2373,7 @@ __skb_to_sgvec(struct sk_buff *skb, struct scatterlist *sg, int offset, int len)
|
||||
for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) {
|
||||
int end;
|
||||
|
||||
BUG_TRAP(start <= offset + len);
|
||||
WARN_ON(start > offset + len);
|
||||
|
||||
end = start + skb_shinfo(skb)->frags[i].size;
|
||||
if ((copy = end - offset) > 0) {
|
||||
@ -2397,7 +2397,7 @@ __skb_to_sgvec(struct sk_buff *skb, struct scatterlist *sg, int offset, int len)
|
||||
for (; list; list = list->next) {
|
||||
int end;
|
||||
|
||||
BUG_TRAP(start <= offset + len);
|
||||
WARN_ON(start > offset + len);
|
||||
|
||||
end = start + list->len;
|
||||
if ((copy = end - offset) > 0) {
|
||||
|
@ -192,13 +192,13 @@ void sk_stream_kill_queues(struct sock *sk)
|
||||
__skb_queue_purge(&sk->sk_error_queue);
|
||||
|
||||
/* Next, the write queue. */
|
||||
BUG_TRAP(skb_queue_empty(&sk->sk_write_queue));
|
||||
WARN_ON(!skb_queue_empty(&sk->sk_write_queue));
|
||||
|
||||
/* Account for returned memory. */
|
||||
sk_mem_reclaim(sk);
|
||||
|
||||
BUG_TRAP(!sk->sk_wmem_queued);
|
||||
BUG_TRAP(!sk->sk_forward_alloc);
|
||||
WARN_ON(sk->sk_wmem_queued);
|
||||
WARN_ON(sk->sk_forward_alloc);
|
||||
|
||||
/* It is _impossible_ for the backlog to contain anything
|
||||
* when we get here. All user references to this socket
|
||||
|
@ -27,7 +27,6 @@
|
||||
|
||||
#include <linux/dmaengine.h>
|
||||
#include <linux/socket.h>
|
||||
#include <linux/rtnetlink.h> /* for BUG_TRAP */
|
||||
#include <net/tcp.h>
|
||||
#include <net/netdma.h>
|
||||
|
||||
@ -72,7 +71,7 @@ int dma_skb_copy_datagram_iovec(struct dma_chan *chan,
|
||||
for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) {
|
||||
int end;
|
||||
|
||||
BUG_TRAP(start <= offset + len);
|
||||
WARN_ON(start > offset + len);
|
||||
|
||||
end = start + skb_shinfo(skb)->frags[i].size;
|
||||
copy = end - offset;
|
||||
@ -101,7 +100,7 @@ int dma_skb_copy_datagram_iovec(struct dma_chan *chan,
|
||||
for (; list; list = list->next) {
|
||||
int end;
|
||||
|
||||
BUG_TRAP(start <= offset + len);
|
||||
WARN_ON(start > offset + len);
|
||||
|
||||
end = start + list->len;
|
||||
copy = end - offset;
|
||||
|
@ -164,7 +164,7 @@ static inline bool dccp_loss_free(const u64 s1, const u64 s2, const u64 ndp)
|
||||
{
|
||||
s64 delta = dccp_delta_seqno(s1, s2);
|
||||
|
||||
BUG_TRAP(delta >= 0);
|
||||
WARN_ON(delta < 0);
|
||||
return (u64)delta <= ndp + 1;
|
||||
}
|
||||
|
||||
|
@ -413,7 +413,7 @@ static int dccp_rcv_request_sent_state_process(struct sock *sk,
|
||||
|
||||
/* Stop the REQUEST timer */
|
||||
inet_csk_clear_xmit_timer(sk, ICSK_TIME_RETRANS);
|
||||
BUG_TRAP(sk->sk_send_head != NULL);
|
||||
WARN_ON(sk->sk_send_head == NULL);
|
||||
__kfree_skb(sk->sk_send_head);
|
||||
sk->sk_send_head = NULL;
|
||||
|
||||
|
@ -283,7 +283,7 @@ static void dccp_v4_err(struct sk_buff *skb, u32 info)
|
||||
* ICMPs are not backlogged, hence we cannot get an established
|
||||
* socket here.
|
||||
*/
|
||||
BUG_TRAP(!req->sk);
|
||||
WARN_ON(req->sk);
|
||||
|
||||
if (seq != dccp_rsk(req)->dreq_iss) {
|
||||
NET_INC_STATS_BH(net, LINUX_MIB_OUTOFWINDOWICMPS);
|
||||
|
@ -186,7 +186,7 @@ static void dccp_v6_err(struct sk_buff *skb, struct inet6_skb_parm *opt,
|
||||
* ICMPs are not backlogged, hence we cannot get an established
|
||||
* socket here.
|
||||
*/
|
||||
BUG_TRAP(req->sk == NULL);
|
||||
WARN_ON(req->sk != NULL);
|
||||
|
||||
if (seq != dccp_rsk(req)->dreq_iss) {
|
||||
NET_INC_STATS_BH(net, LINUX_MIB_OUTOFWINDOWICMPS);
|
||||
|
@ -327,7 +327,7 @@ int dccp_disconnect(struct sock *sk, int flags)
|
||||
inet_csk_delack_init(sk);
|
||||
__sk_dst_reset(sk);
|
||||
|
||||
BUG_TRAP(!inet->num || icsk->icsk_bind_hash);
|
||||
WARN_ON(inet->num && !icsk->icsk_bind_hash);
|
||||
|
||||
sk->sk_error_report(sk);
|
||||
return err;
|
||||
@ -981,7 +981,7 @@ adjudge_to_death:
|
||||
*/
|
||||
local_bh_disable();
|
||||
bh_lock_sock(sk);
|
||||
BUG_TRAP(!sock_owned_by_user(sk));
|
||||
WARN_ON(sock_owned_by_user(sk));
|
||||
|
||||
/* Have we already been destroyed by a softirq or backlog? */
|
||||
if (state != DCCP_CLOSED && sk->sk_state == DCCP_CLOSED)
|
||||
|
@ -106,7 +106,7 @@ static void dccp_retransmit_timer(struct sock *sk)
|
||||
* -- Acks in client-PARTOPEN state (sec. 8.1.5)
|
||||
* -- CloseReq in server-CLOSEREQ state (sec. 8.3)
|
||||
* -- Close in node-CLOSING state (sec. 8.3) */
|
||||
BUG_TRAP(sk->sk_send_head != NULL);
|
||||
WARN_ON(sk->sk_send_head == NULL);
|
||||
|
||||
/*
|
||||
* More than than 4MSL (8 minutes) has passed, a RESET(aborted) was
|
||||
|
@ -148,10 +148,10 @@ void inet_sock_destruct(struct sock *sk)
|
||||
return;
|
||||
}
|
||||
|
||||
BUG_TRAP(!atomic_read(&sk->sk_rmem_alloc));
|
||||
BUG_TRAP(!atomic_read(&sk->sk_wmem_alloc));
|
||||
BUG_TRAP(!sk->sk_wmem_queued);
|
||||
BUG_TRAP(!sk->sk_forward_alloc);
|
||||
WARN_ON(atomic_read(&sk->sk_rmem_alloc));
|
||||
WARN_ON(atomic_read(&sk->sk_wmem_alloc));
|
||||
WARN_ON(sk->sk_wmem_queued);
|
||||
WARN_ON(sk->sk_forward_alloc);
|
||||
|
||||
kfree(inet->opt);
|
||||
dst_release(sk->sk_dst_cache);
|
||||
@ -338,7 +338,7 @@ lookup_protocol:
|
||||
answer_flags = answer->flags;
|
||||
rcu_read_unlock();
|
||||
|
||||
BUG_TRAP(answer_prot->slab != NULL);
|
||||
WARN_ON(answer_prot->slab == NULL);
|
||||
|
||||
err = -ENOBUFS;
|
||||
sk = sk_alloc(net, PF_INET, GFP_KERNEL, answer_prot);
|
||||
@ -658,8 +658,8 @@ int inet_accept(struct socket *sock, struct socket *newsock, int flags)
|
||||
|
||||
lock_sock(sk2);
|
||||
|
||||
BUG_TRAP((1 << sk2->sk_state) &
|
||||
(TCPF_ESTABLISHED | TCPF_CLOSE_WAIT | TCPF_CLOSE));
|
||||
WARN_ON(!((1 << sk2->sk_state) &
|
||||
(TCPF_ESTABLISHED | TCPF_CLOSE_WAIT | TCPF_CLOSE)));
|
||||
|
||||
sock_graft(sk2, newsock);
|
||||
|
||||
|
@ -138,8 +138,8 @@ void in_dev_finish_destroy(struct in_device *idev)
|
||||
{
|
||||
struct net_device *dev = idev->dev;
|
||||
|
||||
BUG_TRAP(!idev->ifa_list);
|
||||
BUG_TRAP(!idev->mc_list);
|
||||
WARN_ON(idev->ifa_list);
|
||||
WARN_ON(idev->mc_list);
|
||||
#ifdef NET_REFCNT_DEBUG
|
||||
printk(KERN_DEBUG "in_dev_finish_destroy: %p=%s\n",
|
||||
idev, dev ? dev->name : "NIL");
|
||||
@ -399,7 +399,7 @@ static int inet_set_ifa(struct net_device *dev, struct in_ifaddr *ifa)
|
||||
}
|
||||
ipv4_devconf_setall(in_dev);
|
||||
if (ifa->ifa_dev != in_dev) {
|
||||
BUG_TRAP(!ifa->ifa_dev);
|
||||
WARN_ON(ifa->ifa_dev);
|
||||
in_dev_hold(in_dev);
|
||||
ifa->ifa_dev = in_dev;
|
||||
}
|
||||
|
@ -167,7 +167,7 @@ tb_not_found:
|
||||
success:
|
||||
if (!inet_csk(sk)->icsk_bind_hash)
|
||||
inet_bind_hash(sk, tb, snum);
|
||||
BUG_TRAP(inet_csk(sk)->icsk_bind_hash == tb);
|
||||
WARN_ON(inet_csk(sk)->icsk_bind_hash != tb);
|
||||
ret = 0;
|
||||
|
||||
fail_unlock:
|
||||
@ -260,7 +260,7 @@ struct sock *inet_csk_accept(struct sock *sk, int flags, int *err)
|
||||
}
|
||||
|
||||
newsk = reqsk_queue_get_child(&icsk->icsk_accept_queue, sk);
|
||||
BUG_TRAP(newsk->sk_state != TCP_SYN_RECV);
|
||||
WARN_ON(newsk->sk_state == TCP_SYN_RECV);
|
||||
out:
|
||||
release_sock(sk);
|
||||
return newsk;
|
||||
@ -386,7 +386,7 @@ struct request_sock *inet_csk_search_req(const struct sock *sk,
|
||||
ireq->rmt_addr == raddr &&
|
||||
ireq->loc_addr == laddr &&
|
||||
AF_INET_FAMILY(req->rsk_ops->family)) {
|
||||
BUG_TRAP(!req->sk);
|
||||
WARN_ON(req->sk);
|
||||
*prevp = prev;
|
||||
break;
|
||||
}
|
||||
@ -539,14 +539,14 @@ EXPORT_SYMBOL_GPL(inet_csk_clone);
|
||||
*/
|
||||
void inet_csk_destroy_sock(struct sock *sk)
|
||||
{
|
||||
BUG_TRAP(sk->sk_state == TCP_CLOSE);
|
||||
BUG_TRAP(sock_flag(sk, SOCK_DEAD));
|
||||
WARN_ON(sk->sk_state != TCP_CLOSE);
|
||||
WARN_ON(!sock_flag(sk, SOCK_DEAD));
|
||||
|
||||
/* It cannot be in hash table! */
|
||||
BUG_TRAP(sk_unhashed(sk));
|
||||
WARN_ON(!sk_unhashed(sk));
|
||||
|
||||
/* If it has not 0 inet_sk(sk)->num, it must be bound */
|
||||
BUG_TRAP(!inet_sk(sk)->num || inet_csk(sk)->icsk_bind_hash);
|
||||
WARN_ON(inet_sk(sk)->num && !inet_csk(sk)->icsk_bind_hash);
|
||||
|
||||
sk->sk_prot->destroy(sk);
|
||||
|
||||
@ -629,7 +629,7 @@ void inet_csk_listen_stop(struct sock *sk)
|
||||
|
||||
local_bh_disable();
|
||||
bh_lock_sock(child);
|
||||
BUG_TRAP(!sock_owned_by_user(child));
|
||||
WARN_ON(sock_owned_by_user(child));
|
||||
sock_hold(child);
|
||||
|
||||
sk->sk_prot->disconnect(child, O_NONBLOCK);
|
||||
@ -647,7 +647,7 @@ void inet_csk_listen_stop(struct sock *sk)
|
||||
sk_acceptq_removed(sk);
|
||||
__reqsk_free(req);
|
||||
}
|
||||
BUG_TRAP(!sk->sk_ack_backlog);
|
||||
WARN_ON(sk->sk_ack_backlog);
|
||||
}
|
||||
|
||||
EXPORT_SYMBOL_GPL(inet_csk_listen_stop);
|
||||
|
@ -134,8 +134,8 @@ void inet_frag_destroy(struct inet_frag_queue *q, struct inet_frags *f,
|
||||
struct sk_buff *fp;
|
||||
struct netns_frags *nf;
|
||||
|
||||
BUG_TRAP(q->last_in & INET_FRAG_COMPLETE);
|
||||
BUG_TRAP(del_timer(&q->timer) == 0);
|
||||
WARN_ON(!(q->last_in & INET_FRAG_COMPLETE));
|
||||
WARN_ON(del_timer(&q->timer) != 0);
|
||||
|
||||
/* Release all fragment data. */
|
||||
fp = q->fragments;
|
||||
|
@ -305,7 +305,7 @@ unique:
|
||||
inet->num = lport;
|
||||
inet->sport = htons(lport);
|
||||
sk->sk_hash = hash;
|
||||
BUG_TRAP(sk_unhashed(sk));
|
||||
WARN_ON(!sk_unhashed(sk));
|
||||
__sk_add_node(sk, &head->chain);
|
||||
sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1);
|
||||
write_unlock(lock);
|
||||
@ -342,7 +342,7 @@ void __inet_hash_nolisten(struct sock *sk)
|
||||
rwlock_t *lock;
|
||||
struct inet_ehash_bucket *head;
|
||||
|
||||
BUG_TRAP(sk_unhashed(sk));
|
||||
WARN_ON(!sk_unhashed(sk));
|
||||
|
||||
sk->sk_hash = inet_sk_ehashfn(sk);
|
||||
head = inet_ehash_bucket(hashinfo, sk->sk_hash);
|
||||
@ -367,7 +367,7 @@ static void __inet_hash(struct sock *sk)
|
||||
return;
|
||||
}
|
||||
|
||||
BUG_TRAP(sk_unhashed(sk));
|
||||
WARN_ON(!sk_unhashed(sk));
|
||||
list = &hashinfo->listening_hash[inet_sk_listen_hashfn(sk)];
|
||||
lock = &hashinfo->lhash_lock;
|
||||
|
||||
@ -450,7 +450,7 @@ int __inet_hash_connect(struct inet_timewait_death_row *death_row,
|
||||
*/
|
||||
inet_bind_bucket_for_each(tb, node, &head->chain) {
|
||||
if (tb->ib_net == net && tb->port == port) {
|
||||
BUG_TRAP(!hlist_empty(&tb->owners));
|
||||
WARN_ON(hlist_empty(&tb->owners));
|
||||
if (tb->fastreuse >= 0)
|
||||
goto next_port;
|
||||
if (!check_established(death_row, sk,
|
||||
|
@ -86,7 +86,7 @@ void __inet_twsk_hashdance(struct inet_timewait_sock *tw, struct sock *sk,
|
||||
hashinfo->bhash_size)];
|
||||
spin_lock(&bhead->lock);
|
||||
tw->tw_tb = icsk->icsk_bind_hash;
|
||||
BUG_TRAP(icsk->icsk_bind_hash);
|
||||
WARN_ON(!icsk->icsk_bind_hash);
|
||||
inet_twsk_add_bind_node(tw, &tw->tw_tb->owners);
|
||||
spin_unlock(&bhead->lock);
|
||||
|
||||
|
@ -488,8 +488,8 @@ static int ip_frag_reasm(struct ipq *qp, struct sk_buff *prev,
|
||||
qp->q.fragments = head;
|
||||
}
|
||||
|
||||
BUG_TRAP(head != NULL);
|
||||
BUG_TRAP(FRAG_CB(head)->offset == 0);
|
||||
WARN_ON(head == NULL);
|
||||
WARN_ON(FRAG_CB(head)->offset != 0);
|
||||
|
||||
/* Allocate a new buffer for the datagram. */
|
||||
ihlen = ip_hdrlen(head);
|
||||
|
@ -118,7 +118,7 @@ static int ip_dev_loopback_xmit(struct sk_buff *newskb)
|
||||
__skb_pull(newskb, skb_network_offset(newskb));
|
||||
newskb->pkt_type = PACKET_LOOPBACK;
|
||||
newskb->ip_summed = CHECKSUM_UNNECESSARY;
|
||||
BUG_TRAP(newskb->dst);
|
||||
WARN_ON(!newskb->dst);
|
||||
netif_rx(newskb);
|
||||
return 0;
|
||||
}
|
||||
|
@ -55,32 +55,53 @@ static struct xt_table packet_filter = {
|
||||
};
|
||||
|
||||
/* The work comes in here from netfilter.c */
|
||||
static unsigned int arpt_hook(unsigned int hook,
|
||||
struct sk_buff *skb,
|
||||
const struct net_device *in,
|
||||
const struct net_device *out,
|
||||
int (*okfn)(struct sk_buff *))
|
||||
static unsigned int arpt_in_hook(unsigned int hook,
|
||||
struct sk_buff *skb,
|
||||
const struct net_device *in,
|
||||
const struct net_device *out,
|
||||
int (*okfn)(struct sk_buff *))
|
||||
{
|
||||
return arpt_do_table(skb, hook, in, out, init_net.ipv4.arptable_filter);
|
||||
return arpt_do_table(skb, hook, in, out,
|
||||
dev_net(in)->ipv4.arptable_filter);
|
||||
}
|
||||
|
||||
static unsigned int arpt_out_hook(unsigned int hook,
|
||||
struct sk_buff *skb,
|
||||
const struct net_device *in,
|
||||
const struct net_device *out,
|
||||
int (*okfn)(struct sk_buff *))
|
||||
{
|
||||
return arpt_do_table(skb, hook, in, out,
|
||||
dev_net(out)->ipv4.arptable_filter);
|
||||
}
|
||||
|
||||
static unsigned int arpt_forward_hook(unsigned int hook,
|
||||
struct sk_buff *skb,
|
||||
const struct net_device *in,
|
||||
const struct net_device *out,
|
||||
int (*okfn)(struct sk_buff *))
|
||||
{
|
||||
return arpt_do_table(skb, hook, in, out,
|
||||
dev_net(in)->ipv4.arptable_filter);
|
||||
}
|
||||
|
||||
static struct nf_hook_ops arpt_ops[] __read_mostly = {
|
||||
{
|
||||
.hook = arpt_hook,
|
||||
.hook = arpt_in_hook,
|
||||
.owner = THIS_MODULE,
|
||||
.pf = NF_ARP,
|
||||
.hooknum = NF_ARP_IN,
|
||||
.priority = NF_IP_PRI_FILTER,
|
||||
},
|
||||
{
|
||||
.hook = arpt_hook,
|
||||
.hook = arpt_out_hook,
|
||||
.owner = THIS_MODULE,
|
||||
.pf = NF_ARP,
|
||||
.hooknum = NF_ARP_OUT,
|
||||
.priority = NF_IP_PRI_FILTER,
|
||||
},
|
||||
{
|
||||
.hook = arpt_hook,
|
||||
.hook = arpt_forward_hook,
|
||||
.owner = THIS_MODULE,
|
||||
.pf = NF_ARP,
|
||||
.hooknum = NF_ARP_FORWARD,
|
||||
|
@ -32,7 +32,7 @@ static struct
|
||||
struct ipt_replace repl;
|
||||
struct ipt_standard entries[3];
|
||||
struct ipt_error term;
|
||||
} initial_table __initdata = {
|
||||
} initial_table __net_initdata = {
|
||||
.repl = {
|
||||
.name = "security",
|
||||
.valid_hooks = SECURITY_VALID_HOOKS,
|
||||
|
@ -1502,7 +1502,7 @@ unsigned short ip_rt_frag_needed(struct net *net, struct iphdr *iph,
|
||||
rth->fl.iif != 0 ||
|
||||
dst_metric_locked(&rth->u.dst, RTAX_MTU) ||
|
||||
!net_eq(dev_net(rth->u.dst.dev), net) ||
|
||||
!rt_is_expired(rth))
|
||||
rt_is_expired(rth))
|
||||
continue;
|
||||
|
||||
if (new_mtu < 68 || new_mtu >= old_mtu) {
|
||||
|
@ -299,6 +299,7 @@ struct sock *cookie_v4_check(struct sock *sk, struct sk_buff *skb,
|
||||
ireq->rmt_port = th->source;
|
||||
ireq->loc_addr = ip_hdr(skb)->daddr;
|
||||
ireq->rmt_addr = ip_hdr(skb)->saddr;
|
||||
ireq->ecn_ok = 0;
|
||||
ireq->snd_wscale = tcp_opt.snd_wscale;
|
||||
ireq->rcv_wscale = tcp_opt.rcv_wscale;
|
||||
ireq->sack_ok = tcp_opt.sack_ok;
|
||||
|
@ -1096,7 +1096,7 @@ void tcp_cleanup_rbuf(struct sock *sk, int copied)
|
||||
#if TCP_DEBUG
|
||||
struct sk_buff *skb = skb_peek(&sk->sk_receive_queue);
|
||||
|
||||
BUG_TRAP(!skb || before(tp->copied_seq, TCP_SKB_CB(skb)->end_seq));
|
||||
WARN_ON(skb && !before(tp->copied_seq, TCP_SKB_CB(skb)->end_seq));
|
||||
#endif
|
||||
|
||||
if (inet_csk_ack_scheduled(sk)) {
|
||||
@ -1358,7 +1358,7 @@ int tcp_recvmsg(struct kiocb *iocb, struct sock *sk, struct msghdr *msg,
|
||||
goto found_ok_skb;
|
||||
if (tcp_hdr(skb)->fin)
|
||||
goto found_fin_ok;
|
||||
BUG_TRAP(flags & MSG_PEEK);
|
||||
WARN_ON(!(flags & MSG_PEEK));
|
||||
skb = skb->next;
|
||||
} while (skb != (struct sk_buff *)&sk->sk_receive_queue);
|
||||
|
||||
@ -1421,8 +1421,8 @@ int tcp_recvmsg(struct kiocb *iocb, struct sock *sk, struct msghdr *msg,
|
||||
|
||||
tp->ucopy.len = len;
|
||||
|
||||
BUG_TRAP(tp->copied_seq == tp->rcv_nxt ||
|
||||
(flags & (MSG_PEEK | MSG_TRUNC)));
|
||||
WARN_ON(tp->copied_seq != tp->rcv_nxt &&
|
||||
!(flags & (MSG_PEEK | MSG_TRUNC)));
|
||||
|
||||
/* Ugly... If prequeue is not empty, we have to
|
||||
* process it before releasing socket, otherwise
|
||||
@ -1844,7 +1844,7 @@ adjudge_to_death:
|
||||
*/
|
||||
local_bh_disable();
|
||||
bh_lock_sock(sk);
|
||||
BUG_TRAP(!sock_owned_by_user(sk));
|
||||
WARN_ON(sock_owned_by_user(sk));
|
||||
|
||||
/* Have we already been destroyed by a softirq or backlog? */
|
||||
if (state != TCP_CLOSE && sk->sk_state == TCP_CLOSE)
|
||||
@ -1973,7 +1973,7 @@ int tcp_disconnect(struct sock *sk, int flags)
|
||||
memset(&tp->rx_opt, 0, sizeof(tp->rx_opt));
|
||||
__sk_dst_reset(sk);
|
||||
|
||||
BUG_TRAP(!inet->num || icsk->icsk_bind_hash);
|
||||
WARN_ON(inet->num && !icsk->icsk_bind_hash);
|
||||
|
||||
sk->sk_error_report(sk);
|
||||
return err;
|
||||
|
@ -1629,10 +1629,10 @@ advance_sp:
|
||||
out:
|
||||
|
||||
#if FASTRETRANS_DEBUG > 0
|
||||
BUG_TRAP((int)tp->sacked_out >= 0);
|
||||
BUG_TRAP((int)tp->lost_out >= 0);
|
||||
BUG_TRAP((int)tp->retrans_out >= 0);
|
||||
BUG_TRAP((int)tcp_packets_in_flight(tp) >= 0);
|
||||
WARN_ON((int)tp->sacked_out < 0);
|
||||
WARN_ON((int)tp->lost_out < 0);
|
||||
WARN_ON((int)tp->retrans_out < 0);
|
||||
WARN_ON((int)tcp_packets_in_flight(tp) < 0);
|
||||
#endif
|
||||
return flag;
|
||||
}
|
||||
@ -2181,7 +2181,7 @@ static void tcp_mark_head_lost(struct sock *sk, int packets)
|
||||
int err;
|
||||
unsigned int mss;
|
||||
|
||||
BUG_TRAP(packets <= tp->packets_out);
|
||||
WARN_ON(packets > tp->packets_out);
|
||||
if (tp->lost_skb_hint) {
|
||||
skb = tp->lost_skb_hint;
|
||||
cnt = tp->lost_cnt_hint;
|
||||
@ -2610,7 +2610,7 @@ static void tcp_fastretrans_alert(struct sock *sk, int pkts_acked, int flag)
|
||||
/* E. Check state exit conditions. State can be terminated
|
||||
* when high_seq is ACKed. */
|
||||
if (icsk->icsk_ca_state == TCP_CA_Open) {
|
||||
BUG_TRAP(tp->retrans_out == 0);
|
||||
WARN_ON(tp->retrans_out != 0);
|
||||
tp->retrans_stamp = 0;
|
||||
} else if (!before(tp->snd_una, tp->high_seq)) {
|
||||
switch (icsk->icsk_ca_state) {
|
||||
@ -2972,9 +2972,9 @@ static int tcp_clean_rtx_queue(struct sock *sk, int prior_fackets)
|
||||
}
|
||||
|
||||
#if FASTRETRANS_DEBUG > 0
|
||||
BUG_TRAP((int)tp->sacked_out >= 0);
|
||||
BUG_TRAP((int)tp->lost_out >= 0);
|
||||
BUG_TRAP((int)tp->retrans_out >= 0);
|
||||
WARN_ON((int)tp->sacked_out < 0);
|
||||
WARN_ON((int)tp->lost_out < 0);
|
||||
WARN_ON((int)tp->retrans_out < 0);
|
||||
if (!tp->packets_out && tcp_is_sack(tp)) {
|
||||
icsk = inet_csk(sk);
|
||||
if (tp->lost_out) {
|
||||
@ -3877,7 +3877,7 @@ static void tcp_sack_remove(struct tcp_sock *tp)
|
||||
int i;
|
||||
|
||||
/* RCV.NXT must cover all the block! */
|
||||
BUG_TRAP(!before(tp->rcv_nxt, sp->end_seq));
|
||||
WARN_ON(before(tp->rcv_nxt, sp->end_seq));
|
||||
|
||||
/* Zap this SACK, by moving forward any other SACKS. */
|
||||
for (i=this_sack+1; i < num_sacks; i++)
|
||||
|
@ -418,7 +418,7 @@ void tcp_v4_err(struct sk_buff *skb, u32 info)
|
||||
/* ICMPs are not backlogged, hence we cannot get
|
||||
an established socket here.
|
||||
*/
|
||||
BUG_TRAP(!req->sk);
|
||||
WARN_ON(req->sk);
|
||||
|
||||
if (seq != tcp_rsk(req)->snt_isn) {
|
||||
NET_INC_STATS_BH(net, LINUX_MIB_OUTOFWINDOWICMPS);
|
||||
|
@ -287,7 +287,7 @@ static void tcp_retransmit_timer(struct sock *sk)
|
||||
if (!tp->packets_out)
|
||||
goto out;
|
||||
|
||||
BUG_TRAP(!tcp_write_queue_empty(sk));
|
||||
WARN_ON(tcp_write_queue_empty(sk));
|
||||
|
||||
if (!tp->snd_wnd && !sock_flag(sk, SOCK_DEAD) &&
|
||||
!((1 << sk->sk_state) & (TCPF_SYN_SENT | TCPF_SYN_RECV))) {
|
||||
|
@ -313,8 +313,10 @@ static void in6_dev_finish_destroy_rcu(struct rcu_head *head)
|
||||
void in6_dev_finish_destroy(struct inet6_dev *idev)
|
||||
{
|
||||
struct net_device *dev = idev->dev;
|
||||
BUG_TRAP(idev->addr_list==NULL);
|
||||
BUG_TRAP(idev->mc_list==NULL);
|
||||
|
||||
WARN_ON(idev->addr_list != NULL);
|
||||
WARN_ON(idev->mc_list != NULL);
|
||||
|
||||
#ifdef NET_REFCNT_DEBUG
|
||||
printk(KERN_DEBUG "in6_dev_finish_destroy: %s\n", dev ? dev->name : "NIL");
|
||||
#endif
|
||||
@ -517,8 +519,9 @@ static void addrconf_fixup_forwarding(struct ctl_table *table, int *p, int old)
|
||||
|
||||
void inet6_ifa_finish_destroy(struct inet6_ifaddr *ifp)
|
||||
{
|
||||
BUG_TRAP(ifp->if_next==NULL);
|
||||
BUG_TRAP(ifp->lst_next==NULL);
|
||||
WARN_ON(ifp->if_next != NULL);
|
||||
WARN_ON(ifp->lst_next != NULL);
|
||||
|
||||
#ifdef NET_REFCNT_DEBUG
|
||||
printk(KERN_DEBUG "inet6_ifa_finish_destroy\n");
|
||||
#endif
|
||||
|
@ -150,7 +150,7 @@ lookup_protocol:
|
||||
answer_flags = answer->flags;
|
||||
rcu_read_unlock();
|
||||
|
||||
BUG_TRAP(answer_prot->slab != NULL);
|
||||
WARN_ON(answer_prot->slab == NULL);
|
||||
|
||||
err = -ENOBUFS;
|
||||
sk = sk_alloc(net, PF_INET6, GFP_KERNEL, answer_prot);
|
||||
|
@ -98,7 +98,7 @@ struct request_sock *inet6_csk_search_req(const struct sock *sk,
|
||||
ipv6_addr_equal(&treq->rmt_addr, raddr) &&
|
||||
ipv6_addr_equal(&treq->loc_addr, laddr) &&
|
||||
(!treq->iif || treq->iif == iif)) {
|
||||
BUG_TRAP(req->sk == NULL);
|
||||
WARN_ON(req->sk != NULL);
|
||||
*prevp = prev;
|
||||
return req;
|
||||
}
|
||||
|
@ -28,7 +28,7 @@ void __inet6_hash(struct sock *sk)
|
||||
struct hlist_head *list;
|
||||
rwlock_t *lock;
|
||||
|
||||
BUG_TRAP(sk_unhashed(sk));
|
||||
WARN_ON(!sk_unhashed(sk));
|
||||
|
||||
if (sk->sk_state == TCP_LISTEN) {
|
||||
list = &hashinfo->listening_hash[inet_sk_listen_hashfn(sk)];
|
||||
@ -202,7 +202,7 @@ unique:
|
||||
* in hash table socket with a funny identity. */
|
||||
inet->num = lport;
|
||||
inet->sport = htons(lport);
|
||||
BUG_TRAP(sk_unhashed(sk));
|
||||
WARN_ON(!sk_unhashed(sk));
|
||||
__sk_add_node(sk, &head->chain);
|
||||
sk->sk_hash = hash;
|
||||
sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1);
|
||||
|
@ -287,7 +287,7 @@ static int fib6_dump_node(struct fib6_walker_t *w)
|
||||
w->leaf = rt;
|
||||
return 1;
|
||||
}
|
||||
BUG_TRAP(res!=0);
|
||||
WARN_ON(res == 0);
|
||||
}
|
||||
w->leaf = NULL;
|
||||
return 0;
|
||||
@ -778,7 +778,7 @@ out:
|
||||
pn->leaf = fib6_find_prefix(info->nl_net, pn);
|
||||
#if RT6_DEBUG >= 2
|
||||
if (!pn->leaf) {
|
||||
BUG_TRAP(pn->leaf != NULL);
|
||||
WARN_ON(pn->leaf == NULL);
|
||||
pn->leaf = info->nl_net->ipv6.ip6_null_entry;
|
||||
}
|
||||
#endif
|
||||
@ -942,7 +942,7 @@ struct fib6_node * fib6_locate(struct fib6_node *root,
|
||||
|
||||
#ifdef CONFIG_IPV6_SUBTREES
|
||||
if (src_len) {
|
||||
BUG_TRAP(saddr!=NULL);
|
||||
WARN_ON(saddr == NULL);
|
||||
if (fn && fn->subtree)
|
||||
fn = fib6_locate_1(fn->subtree, saddr, src_len,
|
||||
offsetof(struct rt6_info, rt6i_src));
|
||||
@ -996,9 +996,9 @@ static struct fib6_node *fib6_repair_tree(struct net *net,
|
||||
RT6_TRACE("fixing tree: plen=%d iter=%d\n", fn->fn_bit, iter);
|
||||
iter++;
|
||||
|
||||
BUG_TRAP(!(fn->fn_flags&RTN_RTINFO));
|
||||
BUG_TRAP(!(fn->fn_flags&RTN_TL_ROOT));
|
||||
BUG_TRAP(fn->leaf==NULL);
|
||||
WARN_ON(fn->fn_flags & RTN_RTINFO);
|
||||
WARN_ON(fn->fn_flags & RTN_TL_ROOT);
|
||||
WARN_ON(fn->leaf != NULL);
|
||||
|
||||
children = 0;
|
||||
child = NULL;
|
||||
@ -1014,7 +1014,7 @@ static struct fib6_node *fib6_repair_tree(struct net *net,
|
||||
fn->leaf = fib6_find_prefix(net, fn);
|
||||
#if RT6_DEBUG >= 2
|
||||
if (fn->leaf==NULL) {
|
||||
BUG_TRAP(fn->leaf);
|
||||
WARN_ON(!fn->leaf);
|
||||
fn->leaf = net->ipv6.ip6_null_entry;
|
||||
}
|
||||
#endif
|
||||
@ -1025,16 +1025,17 @@ static struct fib6_node *fib6_repair_tree(struct net *net,
|
||||
pn = fn->parent;
|
||||
#ifdef CONFIG_IPV6_SUBTREES
|
||||
if (FIB6_SUBTREE(pn) == fn) {
|
||||
BUG_TRAP(fn->fn_flags&RTN_ROOT);
|
||||
WARN_ON(!(fn->fn_flags & RTN_ROOT));
|
||||
FIB6_SUBTREE(pn) = NULL;
|
||||
nstate = FWS_L;
|
||||
} else {
|
||||
BUG_TRAP(!(fn->fn_flags&RTN_ROOT));
|
||||
WARN_ON(fn->fn_flags & RTN_ROOT);
|
||||
#endif
|
||||
if (pn->right == fn) pn->right = child;
|
||||
else if (pn->left == fn) pn->left = child;
|
||||
#if RT6_DEBUG >= 2
|
||||
else BUG_TRAP(0);
|
||||
else
|
||||
WARN_ON(1);
|
||||
#endif
|
||||
if (child)
|
||||
child->parent = pn;
|
||||
@ -1154,14 +1155,14 @@ int fib6_del(struct rt6_info *rt, struct nl_info *info)
|
||||
|
||||
#if RT6_DEBUG >= 2
|
||||
if (rt->u.dst.obsolete>0) {
|
||||
BUG_TRAP(fn==NULL);
|
||||
WARN_ON(fn != NULL);
|
||||
return -ENOENT;
|
||||
}
|
||||
#endif
|
||||
if (fn == NULL || rt == net->ipv6.ip6_null_entry)
|
||||
return -ENOENT;
|
||||
|
||||
BUG_TRAP(fn->fn_flags&RTN_RTINFO);
|
||||
WARN_ON(!(fn->fn_flags & RTN_RTINFO));
|
||||
|
||||
if (!(rt->rt6i_flags&RTF_CACHE)) {
|
||||
struct fib6_node *pn = fn;
|
||||
@ -1266,7 +1267,7 @@ static int fib6_walk_continue(struct fib6_walker_t *w)
|
||||
w->node = pn;
|
||||
#ifdef CONFIG_IPV6_SUBTREES
|
||||
if (FIB6_SUBTREE(pn) == fn) {
|
||||
BUG_TRAP(fn->fn_flags&RTN_ROOT);
|
||||
WARN_ON(!(fn->fn_flags & RTN_ROOT));
|
||||
w->state = FWS_L;
|
||||
continue;
|
||||
}
|
||||
@ -1281,7 +1282,7 @@ static int fib6_walk_continue(struct fib6_walker_t *w)
|
||||
continue;
|
||||
}
|
||||
#if RT6_DEBUG >= 2
|
||||
BUG_TRAP(0);
|
||||
WARN_ON(1);
|
||||
#endif
|
||||
}
|
||||
}
|
||||
@ -1323,7 +1324,7 @@ static int fib6_clean_node(struct fib6_walker_t *w)
|
||||
}
|
||||
return 0;
|
||||
}
|
||||
BUG_TRAP(res==0);
|
||||
WARN_ON(res != 0);
|
||||
}
|
||||
w->leaf = rt;
|
||||
return 0;
|
||||
|
@ -116,7 +116,7 @@ static int ip6_dev_loopback_xmit(struct sk_buff *newskb)
|
||||
__skb_pull(newskb, skb_network_offset(newskb));
|
||||
newskb->pkt_type = PACKET_LOOPBACK;
|
||||
newskb->ip_summed = CHECKSUM_UNNECESSARY;
|
||||
BUG_TRAP(newskb->dst);
|
||||
WARN_ON(!newskb->dst);
|
||||
|
||||
netif_rx(newskb);
|
||||
return 0;
|
||||
|
@ -164,8 +164,8 @@ static int mip6_destopt_output(struct xfrm_state *x, struct sk_buff *skb)
|
||||
calc_padlen(sizeof(*dstopt), 6));
|
||||
|
||||
hao->type = IPV6_TLV_HAO;
|
||||
BUILD_BUG_ON(sizeof(*hao) != 18);
|
||||
hao->length = sizeof(*hao) - 2;
|
||||
BUG_TRAP(hao->length == 16);
|
||||
|
||||
len = ((char *)hao - (char *)dstopt) + sizeof(*hao);
|
||||
|
||||
@ -174,7 +174,7 @@ static int mip6_destopt_output(struct xfrm_state *x, struct sk_buff *skb)
|
||||
memcpy(&iph->saddr, x->coaddr, sizeof(iph->saddr));
|
||||
spin_unlock_bh(&x->lock);
|
||||
|
||||
BUG_TRAP(len == x->props.header_len);
|
||||
WARN_ON(len != x->props.header_len);
|
||||
dstopt->hdrlen = (x->props.header_len >> 3) - 1;
|
||||
|
||||
return 0;
|
||||
@ -317,7 +317,7 @@ static int mip6_destopt_init_state(struct xfrm_state *x)
|
||||
x->props.header_len = sizeof(struct ipv6_destopt_hdr) +
|
||||
calc_padlen(sizeof(struct ipv6_destopt_hdr), 6) +
|
||||
sizeof(struct ipv6_destopt_hao);
|
||||
BUG_TRAP(x->props.header_len == 24);
|
||||
WARN_ON(x->props.header_len != 24);
|
||||
|
||||
return 0;
|
||||
}
|
||||
@ -380,7 +380,7 @@ static int mip6_rthdr_output(struct xfrm_state *x, struct sk_buff *skb)
|
||||
rt2->rt_hdr.segments_left = 1;
|
||||
memset(&rt2->reserved, 0, sizeof(rt2->reserved));
|
||||
|
||||
BUG_TRAP(rt2->rt_hdr.hdrlen == 2);
|
||||
WARN_ON(rt2->rt_hdr.hdrlen != 2);
|
||||
|
||||
memcpy(&rt2->addr, &iph->daddr, sizeof(rt2->addr));
|
||||
spin_lock_bh(&x->lock);
|
||||
|
@ -31,7 +31,7 @@ static struct
|
||||
struct ip6t_replace repl;
|
||||
struct ip6t_standard entries[3];
|
||||
struct ip6t_error term;
|
||||
} initial_table __initdata = {
|
||||
} initial_table __net_initdata = {
|
||||
.repl = {
|
||||
.name = "security",
|
||||
.valid_hooks = SECURITY_VALID_HOOKS,
|
||||
|
@ -416,8 +416,8 @@ nf_ct_frag6_reasm(struct nf_ct_frag6_queue *fq, struct net_device *dev)
|
||||
|
||||
fq_kill(fq);
|
||||
|
||||
BUG_TRAP(head != NULL);
|
||||
BUG_TRAP(NFCT_FRAG6_CB(head)->offset == 0);
|
||||
WARN_ON(head == NULL);
|
||||
WARN_ON(NFCT_FRAG6_CB(head)->offset != 0);
|
||||
|
||||
/* Unfragmented part is taken from the first segment. */
|
||||
payload_len = ((head->data - skb_network_header(head)) -
|
||||
|
@ -473,8 +473,8 @@ static int ip6_frag_reasm(struct frag_queue *fq, struct sk_buff *prev,
|
||||
fq->q.fragments = head;
|
||||
}
|
||||
|
||||
BUG_TRAP(head != NULL);
|
||||
BUG_TRAP(FRAG6_CB(head)->offset == 0);
|
||||
WARN_ON(head == NULL);
|
||||
WARN_ON(FRAG6_CB(head)->offset != 0);
|
||||
|
||||
/* Unfragmented part is taken from the first segment. */
|
||||
payload_len = ((head->data - skb_network_header(head)) -
|
||||
|
@ -223,6 +223,7 @@ struct sock *cookie_v6_check(struct sock *sk, struct sk_buff *skb)
|
||||
|
||||
req->expires = 0UL;
|
||||
req->retrans = 0;
|
||||
ireq->ecn_ok = 0;
|
||||
ireq->snd_wscale = tcp_opt.snd_wscale;
|
||||
ireq->rcv_wscale = tcp_opt.rcv_wscale;
|
||||
ireq->sack_ok = tcp_opt.sack_ok;
|
||||
|
@ -421,7 +421,7 @@ static void tcp_v6_err(struct sk_buff *skb, struct inet6_skb_parm *opt,
|
||||
/* ICMPs are not backlogged, hence we cannot get
|
||||
* an established socket here.
|
||||
*/
|
||||
BUG_TRAP(req->sk == NULL);
|
||||
WARN_ON(req->sk != NULL);
|
||||
|
||||
if (seq != tcp_rsk(req)->snt_isn) {
|
||||
NET_INC_STATS_BH(net, LINUX_MIB_OUTOFWINDOWICMPS);
|
||||
|
@ -96,8 +96,8 @@ static void pfkey_sock_destruct(struct sock *sk)
|
||||
return;
|
||||
}
|
||||
|
||||
BUG_TRAP(!atomic_read(&sk->sk_rmem_alloc));
|
||||
BUG_TRAP(!atomic_read(&sk->sk_wmem_alloc));
|
||||
WARN_ON(atomic_read(&sk->sk_rmem_alloc));
|
||||
WARN_ON(atomic_read(&sk->sk_wmem_alloc));
|
||||
|
||||
atomic_dec(&pfkey_socks_nr);
|
||||
}
|
||||
|
@ -95,7 +95,7 @@ void *__nf_ct_ext_add(struct nf_conn *ct, enum nf_ct_ext_id id, gfp_t gfp)
|
||||
newlen = newoff + t->len;
|
||||
rcu_read_unlock();
|
||||
|
||||
new = krealloc(ct->ext, newlen, gfp);
|
||||
new = __krealloc(ct->ext, newlen, gfp);
|
||||
if (!new)
|
||||
return NULL;
|
||||
|
||||
@ -115,10 +115,10 @@ void *__nf_ct_ext_add(struct nf_conn *ct, enum nf_ct_ext_id id, gfp_t gfp)
|
||||
ct->ext = new;
|
||||
}
|
||||
|
||||
ct->ext->offset[id] = newoff;
|
||||
ct->ext->len = newlen;
|
||||
memset((void *)ct->ext + newoff, 0, newlen - newoff);
|
||||
return (void *)ct->ext + newoff;
|
||||
new->offset[id] = newoff;
|
||||
new->len = newlen;
|
||||
memset((void *)new + newoff, 0, newlen - newoff);
|
||||
return (void *)new + newoff;
|
||||
}
|
||||
EXPORT_SYMBOL(__nf_ct_ext_add);
|
||||
|
||||
|
@ -158,9 +158,10 @@ static void netlink_sock_destruct(struct sock *sk)
|
||||
printk(KERN_ERR "Freeing alive netlink socket %p\n", sk);
|
||||
return;
|
||||
}
|
||||
BUG_TRAP(!atomic_read(&sk->sk_rmem_alloc));
|
||||
BUG_TRAP(!atomic_read(&sk->sk_wmem_alloc));
|
||||
BUG_TRAP(!nlk_sk(sk)->groups);
|
||||
|
||||
WARN_ON(atomic_read(&sk->sk_rmem_alloc));
|
||||
WARN_ON(atomic_read(&sk->sk_wmem_alloc));
|
||||
WARN_ON(nlk_sk(sk)->groups);
|
||||
}
|
||||
|
||||
/* This lock without WQ_FLAG_EXCLUSIVE is good on UP and it is _very_ bad on
|
||||
|
@ -260,8 +260,8 @@ static inline struct packet_sock *pkt_sk(struct sock *sk)
|
||||
|
||||
static void packet_sock_destruct(struct sock *sk)
|
||||
{
|
||||
BUG_TRAP(!atomic_read(&sk->sk_rmem_alloc));
|
||||
BUG_TRAP(!atomic_read(&sk->sk_wmem_alloc));
|
||||
WARN_ON(atomic_read(&sk->sk_rmem_alloc));
|
||||
WARN_ON(atomic_read(&sk->sk_wmem_alloc));
|
||||
|
||||
if (!sock_flag(sk, SOCK_DEAD)) {
|
||||
printk("Attempt to release alive packet socket: %p\n", sk);
|
||||
|
@ -660,9 +660,9 @@ static void rxrpc_sock_destructor(struct sock *sk)
|
||||
|
||||
rxrpc_purge_queue(&sk->sk_receive_queue);
|
||||
|
||||
BUG_TRAP(!atomic_read(&sk->sk_wmem_alloc));
|
||||
BUG_TRAP(sk_unhashed(sk));
|
||||
BUG_TRAP(!sk->sk_socket);
|
||||
WARN_ON(atomic_read(&sk->sk_wmem_alloc));
|
||||
WARN_ON(!sk_unhashed(sk));
|
||||
WARN_ON(sk->sk_socket);
|
||||
|
||||
if (!sock_flag(sk, SOCK_DEAD)) {
|
||||
printk("Attempt to release alive rxrpc socket: %p\n", sk);
|
||||
|
@ -41,7 +41,7 @@ void tcf_hash_destroy(struct tcf_common *p, struct tcf_hashinfo *hinfo)
|
||||
return;
|
||||
}
|
||||
}
|
||||
BUG_TRAP(0);
|
||||
WARN_ON(1);
|
||||
}
|
||||
EXPORT_SYMBOL(tcf_hash_destroy);
|
||||
|
||||
|
@ -116,7 +116,7 @@ static void tcf_police_destroy(struct tcf_police *p)
|
||||
return;
|
||||
}
|
||||
}
|
||||
BUG_TRAP(0);
|
||||
WARN_ON(1);
|
||||
}
|
||||
|
||||
static const struct nla_policy police_policy[TCA_POLICE_MAX + 1] = {
|
||||
|
@ -345,7 +345,7 @@ static int u32_delete_key(struct tcf_proto *tp, struct tc_u_knode* key)
|
||||
}
|
||||
}
|
||||
}
|
||||
BUG_TRAP(0);
|
||||
WARN_ON(1);
|
||||
return 0;
|
||||
}
|
||||
|
||||
@ -368,7 +368,7 @@ static int u32_destroy_hnode(struct tcf_proto *tp, struct tc_u_hnode *ht)
|
||||
struct tc_u_common *tp_c = tp->data;
|
||||
struct tc_u_hnode **hn;
|
||||
|
||||
BUG_TRAP(!ht->refcnt);
|
||||
WARN_ON(ht->refcnt);
|
||||
|
||||
u32_clear_hnode(tp, ht);
|
||||
|
||||
@ -380,7 +380,7 @@ static int u32_destroy_hnode(struct tcf_proto *tp, struct tc_u_hnode *ht)
|
||||
}
|
||||
}
|
||||
|
||||
BUG_TRAP(0);
|
||||
WARN_ON(1);
|
||||
return -ENOENT;
|
||||
}
|
||||
|
||||
@ -389,7 +389,7 @@ static void u32_destroy(struct tcf_proto *tp)
|
||||
struct tc_u_common *tp_c = tp->data;
|
||||
struct tc_u_hnode *root_ht = xchg(&tp->root, NULL);
|
||||
|
||||
BUG_TRAP(root_ht != NULL);
|
||||
WARN_ON(root_ht == NULL);
|
||||
|
||||
if (root_ht && --root_ht->refcnt == 0)
|
||||
u32_destroy_hnode(tp, root_ht);
|
||||
@ -407,7 +407,7 @@ static void u32_destroy(struct tcf_proto *tp)
|
||||
while ((ht = tp_c->hlist) != NULL) {
|
||||
tp_c->hlist = ht->next;
|
||||
|
||||
BUG_TRAP(ht->refcnt == 0);
|
||||
WARN_ON(ht->refcnt != 0);
|
||||
|
||||
kfree(ht);
|
||||
}
|
||||
|
@ -1175,7 +1175,7 @@ static void cbq_unlink_class(struct cbq_class *this)
|
||||
this->tparent->children = NULL;
|
||||
}
|
||||
} else {
|
||||
BUG_TRAP(this->sibling == this);
|
||||
WARN_ON(this->sibling != this);
|
||||
}
|
||||
}
|
||||
|
||||
@ -1699,7 +1699,7 @@ static void cbq_destroy_class(struct Qdisc *sch, struct cbq_class *cl)
|
||||
{
|
||||
struct cbq_sched_data *q = qdisc_priv(sch);
|
||||
|
||||
BUG_TRAP(!cl->filters);
|
||||
WARN_ON(cl->filters);
|
||||
|
||||
tcf_destroy_chain(&cl->filter_list);
|
||||
qdisc_destroy(cl->q);
|
||||
|
@ -746,5 +746,5 @@ void dev_shutdown(struct net_device *dev)
|
||||
{
|
||||
netdev_for_each_tx_queue(dev, shutdown_scheduler_queue, &noop_qdisc);
|
||||
shutdown_scheduler_queue(dev, &dev->rx_queue, NULL);
|
||||
BUG_TRAP(!timer_pending(&dev->watchdog_timer));
|
||||
WARN_ON(timer_pending(&dev->watchdog_timer));
|
||||
}
|
||||
|
@ -524,7 +524,7 @@ htb_change_class_mode(struct htb_sched *q, struct htb_class *cl, long *diff)
|
||||
*/
|
||||
static inline void htb_activate(struct htb_sched *q, struct htb_class *cl)
|
||||
{
|
||||
BUG_TRAP(!cl->level && cl->un.leaf.q && cl->un.leaf.q->q.qlen);
|
||||
WARN_ON(cl->level || !cl->un.leaf.q || !cl->un.leaf.q->q.qlen);
|
||||
|
||||
if (!cl->prio_activity) {
|
||||
cl->prio_activity = 1 << (cl->un.leaf.aprio = cl->un.leaf.prio);
|
||||
@ -542,7 +542,7 @@ static inline void htb_activate(struct htb_sched *q, struct htb_class *cl)
|
||||
*/
|
||||
static inline void htb_deactivate(struct htb_sched *q, struct htb_class *cl)
|
||||
{
|
||||
BUG_TRAP(cl->prio_activity);
|
||||
WARN_ON(!cl->prio_activity);
|
||||
|
||||
htb_deactivate_prios(q, cl);
|
||||
cl->prio_activity = 0;
|
||||
@ -757,7 +757,7 @@ static struct htb_class *htb_lookup_leaf(struct rb_root *tree, int prio,
|
||||
u32 *pid;
|
||||
} stk[TC_HTB_MAXDEPTH], *sp = stk;
|
||||
|
||||
BUG_TRAP(tree->rb_node);
|
||||
WARN_ON(!tree->rb_node);
|
||||
sp->root = tree->rb_node;
|
||||
sp->pptr = pptr;
|
||||
sp->pid = pid;
|
||||
@ -777,7 +777,7 @@ static struct htb_class *htb_lookup_leaf(struct rb_root *tree, int prio,
|
||||
*sp->pptr = (*sp->pptr)->rb_left;
|
||||
if (sp > stk) {
|
||||
sp--;
|
||||
BUG_TRAP(*sp->pptr);
|
||||
WARN_ON(!*sp->pptr);
|
||||
if (!*sp->pptr)
|
||||
return NULL;
|
||||
htb_next_rb_node(sp->pptr);
|
||||
@ -792,7 +792,7 @@ static struct htb_class *htb_lookup_leaf(struct rb_root *tree, int prio,
|
||||
sp->pid = cl->un.inner.last_ptr_id + prio;
|
||||
}
|
||||
}
|
||||
BUG_TRAP(0);
|
||||
WARN_ON(1);
|
||||
return NULL;
|
||||
}
|
||||
|
||||
@ -810,7 +810,7 @@ static struct sk_buff *htb_dequeue_tree(struct htb_sched *q, int prio,
|
||||
|
||||
do {
|
||||
next:
|
||||
BUG_TRAP(cl);
|
||||
WARN_ON(!cl);
|
||||
if (!cl)
|
||||
return NULL;
|
||||
|
||||
@ -1185,7 +1185,7 @@ static void htb_parent_to_leaf(struct htb_sched *q, struct htb_class *cl,
|
||||
{
|
||||
struct htb_class *parent = cl->parent;
|
||||
|
||||
BUG_TRAP(!cl->level && cl->un.leaf.q && !cl->prio_activity);
|
||||
WARN_ON(cl->level || !cl->un.leaf.q || cl->prio_activity);
|
||||
|
||||
if (parent->cmode != HTB_CAN_SEND)
|
||||
htb_safe_rb_erase(&parent->pq_node, q->wait_pq + parent->level);
|
||||
@ -1205,7 +1205,7 @@ static void htb_parent_to_leaf(struct htb_sched *q, struct htb_class *cl,
|
||||
static void htb_destroy_class(struct Qdisc *sch, struct htb_class *cl)
|
||||
{
|
||||
if (!cl->level) {
|
||||
BUG_TRAP(cl->un.leaf.q);
|
||||
WARN_ON(!cl->un.leaf.q);
|
||||
qdisc_destroy(cl->un.leaf.q);
|
||||
}
|
||||
gen_kill_estimator(&cl->bstats, &cl->rate_est);
|
||||
|
@ -536,14 +536,7 @@ static int sfq_dump(struct Qdisc *sch, struct sk_buff *skb)
|
||||
|
||||
opt.limit = q->limit;
|
||||
opt.divisor = SFQ_HASH_DIVISOR;
|
||||
opt.flows = 0;
|
||||
if (q->tail != SFQ_DEPTH) {
|
||||
unsigned int i;
|
||||
|
||||
for (i = 0; i < SFQ_HASH_DIVISOR; i++)
|
||||
if (q->ht[i] != SFQ_DEPTH)
|
||||
opt.flows++;
|
||||
}
|
||||
opt.flows = q->limit;
|
||||
|
||||
NLA_PUT(skb, TCA_OPTIONS, sizeof(opt), &opt);
|
||||
|
||||
|
@ -464,7 +464,7 @@ static void sctp_association_destroy(struct sctp_association *asoc)
|
||||
spin_unlock_bh(&sctp_assocs_id_lock);
|
||||
}
|
||||
|
||||
BUG_TRAP(!atomic_read(&asoc->rmem_alloc));
|
||||
WARN_ON(atomic_read(&asoc->rmem_alloc));
|
||||
|
||||
if (asoc->base.malloced) {
|
||||
kfree(asoc);
|
||||
|
@ -227,7 +227,7 @@ static void __unix_remove_socket(struct sock *sk)
|
||||
|
||||
static void __unix_insert_socket(struct hlist_head *list, struct sock *sk)
|
||||
{
|
||||
BUG_TRAP(sk_unhashed(sk));
|
||||
WARN_ON(!sk_unhashed(sk));
|
||||
sk_add_node(sk, list);
|
||||
}
|
||||
|
||||
@ -350,9 +350,9 @@ static void unix_sock_destructor(struct sock *sk)
|
||||
|
||||
skb_queue_purge(&sk->sk_receive_queue);
|
||||
|
||||
BUG_TRAP(!atomic_read(&sk->sk_wmem_alloc));
|
||||
BUG_TRAP(sk_unhashed(sk));
|
||||
BUG_TRAP(!sk->sk_socket);
|
||||
WARN_ON(atomic_read(&sk->sk_wmem_alloc));
|
||||
WARN_ON(!sk_unhashed(sk));
|
||||
WARN_ON(sk->sk_socket);
|
||||
if (!sock_flag(sk, SOCK_DEAD)) {
|
||||
printk("Attempt to release alive unix socket: %p\n", sk);
|
||||
return;
|
||||
|
@ -718,7 +718,7 @@ int skb_icv_walk(const struct sk_buff *skb, struct hash_desc *desc,
|
||||
for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) {
|
||||
int end;
|
||||
|
||||
BUG_TRAP(start <= offset + len);
|
||||
WARN_ON(start > offset + len);
|
||||
|
||||
end = start + skb_shinfo(skb)->frags[i].size;
|
||||
if ((copy = end - offset) > 0) {
|
||||
@ -748,7 +748,7 @@ int skb_icv_walk(const struct sk_buff *skb, struct hash_desc *desc,
|
||||
for (; list; list = list->next) {
|
||||
int end;
|
||||
|
||||
BUG_TRAP(start <= offset + len);
|
||||
WARN_ON(start > offset + len);
|
||||
|
||||
end = start + list->len;
|
||||
if ((copy = end - offset) > 0) {
|
||||
|
@ -22,7 +22,6 @@
|
||||
#include <linux/module.h>
|
||||
#include <linux/mutex.h>
|
||||
#include <linux/percpu.h>
|
||||
#include <linux/rtnetlink.h>
|
||||
#include <linux/smp.h>
|
||||
#include <linux/vmalloc.h>
|
||||
#include <net/ip.h>
|
||||
@ -251,7 +250,7 @@ static void ipcomp_free_tfms(struct crypto_comp **tfms)
|
||||
break;
|
||||
}
|
||||
|
||||
BUG_TRAP(pos);
|
||||
WARN_ON(!pos);
|
||||
|
||||
if (--pos->users)
|
||||
return;
|
||||
|
@ -538,7 +538,7 @@ EXPORT_SYMBOL(xfrm_state_alloc);
|
||||
|
||||
void __xfrm_state_destroy(struct xfrm_state *x)
|
||||
{
|
||||
BUG_TRAP(x->km.state == XFRM_STATE_DEAD);
|
||||
WARN_ON(x->km.state != XFRM_STATE_DEAD);
|
||||
|
||||
spin_lock_bh(&xfrm_state_lock);
|
||||
list_del(&x->all);
|
||||
|
@ -5654,27 +5654,20 @@ static struct nf_hook_ops selinux_ipv6_ops[] = {
|
||||
static int __init selinux_nf_ip_init(void)
|
||||
{
|
||||
int err = 0;
|
||||
u32 iter;
|
||||
|
||||
if (!selinux_enabled)
|
||||
goto out;
|
||||
|
||||
printk(KERN_DEBUG "SELinux: Registering netfilter hooks\n");
|
||||
|
||||
for (iter = 0; iter < ARRAY_SIZE(selinux_ipv4_ops); iter++) {
|
||||
err = nf_register_hook(&selinux_ipv4_ops[iter]);
|
||||
if (err)
|
||||
panic("SELinux: nf_register_hook for IPv4: error %d\n",
|
||||
err);
|
||||
}
|
||||
err = nf_register_hooks(selinux_ipv4_ops, ARRAY_SIZE(selinux_ipv4_ops));
|
||||
if (err)
|
||||
panic("SELinux: nf_register_hooks for IPv4: error %d\n", err);
|
||||
|
||||
#if defined(CONFIG_IPV6) || defined(CONFIG_IPV6_MODULE)
|
||||
for (iter = 0; iter < ARRAY_SIZE(selinux_ipv6_ops); iter++) {
|
||||
err = nf_register_hook(&selinux_ipv6_ops[iter]);
|
||||
if (err)
|
||||
panic("SELinux: nf_register_hook for IPv6: error %d\n",
|
||||
err);
|
||||
}
|
||||
err = nf_register_hooks(selinux_ipv6_ops, ARRAY_SIZE(selinux_ipv6_ops));
|
||||
if (err)
|
||||
panic("SELinux: nf_register_hooks for IPv6: error %d\n", err);
|
||||
#endif /* IPV6 */
|
||||
|
||||
out:
|
||||
@ -5686,15 +5679,11 @@ __initcall(selinux_nf_ip_init);
|
||||
#ifdef CONFIG_SECURITY_SELINUX_DISABLE
|
||||
static void selinux_nf_ip_exit(void)
|
||||
{
|
||||
u32 iter;
|
||||
|
||||
printk(KERN_DEBUG "SELinux: Unregistering netfilter hooks\n");
|
||||
|
||||
for (iter = 0; iter < ARRAY_SIZE(selinux_ipv4_ops); iter++)
|
||||
nf_unregister_hook(&selinux_ipv4_ops[iter]);
|
||||
nf_unregister_hooks(selinux_ipv4_ops, ARRAY_SIZE(selinux_ipv4_ops));
|
||||
#if defined(CONFIG_IPV6) || defined(CONFIG_IPV6_MODULE)
|
||||
for (iter = 0; iter < ARRAY_SIZE(selinux_ipv6_ops); iter++)
|
||||
nf_unregister_hook(&selinux_ipv6_ops[iter]);
|
||||
nf_unregister_hooks(selinux_ipv6_ops, ARRAY_SIZE(selinux_ipv6_ops));
|
||||
#endif /* IPV6 */
|
||||
}
|
||||
#endif
|
||||
|
Loading…
Reference in New Issue
Block a user