mirror of
https://github.com/torvalds/linux.git
synced 2024-11-14 08:02:07 +00:00
b4bffa4cea
Add missing extacks for common configuration errors. Signed-off-by: Guillaume Nault <gnault@redhat.com> Signed-off-by: David S. Miller <davem@davemloft.net>
805 lines
20 KiB
C
805 lines
20 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/* Bareudp: UDP tunnel encasulation for different Payload types like
|
|
* MPLS, NSH, IP, etc.
|
|
* Copyright (c) 2019 Nokia, Inc.
|
|
* Authors: Martin Varghese, <martin.varghese@nokia.com>
|
|
*/
|
|
|
|
#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
|
|
|
|
#include <linux/kernel.h>
|
|
#include <linux/module.h>
|
|
#include <linux/etherdevice.h>
|
|
#include <linux/hash.h>
|
|
#include <net/dst_metadata.h>
|
|
#include <net/gro_cells.h>
|
|
#include <net/rtnetlink.h>
|
|
#include <net/protocol.h>
|
|
#include <net/ip6_tunnel.h>
|
|
#include <net/ip_tunnels.h>
|
|
#include <net/udp_tunnel.h>
|
|
#include <net/bareudp.h>
|
|
|
|
#define BAREUDP_BASE_HLEN sizeof(struct udphdr)
|
|
#define BAREUDP_IPV4_HLEN (sizeof(struct iphdr) + \
|
|
sizeof(struct udphdr))
|
|
#define BAREUDP_IPV6_HLEN (sizeof(struct ipv6hdr) + \
|
|
sizeof(struct udphdr))
|
|
|
|
static bool log_ecn_error = true;
|
|
module_param(log_ecn_error, bool, 0644);
|
|
MODULE_PARM_DESC(log_ecn_error, "Log packets received with corrupted ECN");
|
|
|
|
/* per-network namespace private data for this module */
|
|
|
|
static unsigned int bareudp_net_id;
|
|
|
|
struct bareudp_net {
|
|
struct list_head bareudp_list;
|
|
};
|
|
|
|
struct bareudp_conf {
|
|
__be16 ethertype;
|
|
__be16 port;
|
|
u16 sport_min;
|
|
bool multi_proto_mode;
|
|
};
|
|
|
|
/* Pseudo network device */
|
|
struct bareudp_dev {
|
|
struct net *net; /* netns for packet i/o */
|
|
struct net_device *dev; /* netdev for bareudp tunnel */
|
|
__be16 ethertype;
|
|
__be16 port;
|
|
u16 sport_min;
|
|
bool multi_proto_mode;
|
|
struct socket __rcu *sock;
|
|
struct list_head next; /* bareudp node on namespace list */
|
|
struct gro_cells gro_cells;
|
|
};
|
|
|
|
static int bareudp_udp_encap_recv(struct sock *sk, struct sk_buff *skb)
|
|
{
|
|
struct metadata_dst *tun_dst = NULL;
|
|
struct bareudp_dev *bareudp;
|
|
unsigned short family;
|
|
unsigned int len;
|
|
__be16 proto;
|
|
void *oiph;
|
|
int err;
|
|
|
|
bareudp = rcu_dereference_sk_user_data(sk);
|
|
if (!bareudp)
|
|
goto drop;
|
|
|
|
if (skb->protocol == htons(ETH_P_IP))
|
|
family = AF_INET;
|
|
else
|
|
family = AF_INET6;
|
|
|
|
if (bareudp->ethertype == htons(ETH_P_IP)) {
|
|
__u8 ipversion;
|
|
|
|
if (skb_copy_bits(skb, BAREUDP_BASE_HLEN, &ipversion,
|
|
sizeof(ipversion))) {
|
|
bareudp->dev->stats.rx_dropped++;
|
|
goto drop;
|
|
}
|
|
ipversion >>= 4;
|
|
|
|
if (ipversion == 4) {
|
|
proto = htons(ETH_P_IP);
|
|
} else if (ipversion == 6 && bareudp->multi_proto_mode) {
|
|
proto = htons(ETH_P_IPV6);
|
|
} else {
|
|
bareudp->dev->stats.rx_dropped++;
|
|
goto drop;
|
|
}
|
|
} else if (bareudp->ethertype == htons(ETH_P_MPLS_UC)) {
|
|
struct iphdr *tunnel_hdr;
|
|
|
|
tunnel_hdr = (struct iphdr *)skb_network_header(skb);
|
|
if (tunnel_hdr->version == 4) {
|
|
if (!ipv4_is_multicast(tunnel_hdr->daddr)) {
|
|
proto = bareudp->ethertype;
|
|
} else if (bareudp->multi_proto_mode &&
|
|
ipv4_is_multicast(tunnel_hdr->daddr)) {
|
|
proto = htons(ETH_P_MPLS_MC);
|
|
} else {
|
|
bareudp->dev->stats.rx_dropped++;
|
|
goto drop;
|
|
}
|
|
} else {
|
|
int addr_type;
|
|
struct ipv6hdr *tunnel_hdr_v6;
|
|
|
|
tunnel_hdr_v6 = (struct ipv6hdr *)skb_network_header(skb);
|
|
addr_type =
|
|
ipv6_addr_type((struct in6_addr *)&tunnel_hdr_v6->daddr);
|
|
if (!(addr_type & IPV6_ADDR_MULTICAST)) {
|
|
proto = bareudp->ethertype;
|
|
} else if (bareudp->multi_proto_mode &&
|
|
(addr_type & IPV6_ADDR_MULTICAST)) {
|
|
proto = htons(ETH_P_MPLS_MC);
|
|
} else {
|
|
bareudp->dev->stats.rx_dropped++;
|
|
goto drop;
|
|
}
|
|
}
|
|
} else {
|
|
proto = bareudp->ethertype;
|
|
}
|
|
|
|
if (iptunnel_pull_header(skb, BAREUDP_BASE_HLEN,
|
|
proto,
|
|
!net_eq(bareudp->net,
|
|
dev_net(bareudp->dev)))) {
|
|
bareudp->dev->stats.rx_dropped++;
|
|
goto drop;
|
|
}
|
|
tun_dst = udp_tun_rx_dst(skb, family, TUNNEL_KEY, 0, 0);
|
|
if (!tun_dst) {
|
|
bareudp->dev->stats.rx_dropped++;
|
|
goto drop;
|
|
}
|
|
skb_dst_set(skb, &tun_dst->dst);
|
|
skb->dev = bareudp->dev;
|
|
oiph = skb_network_header(skb);
|
|
skb_reset_network_header(skb);
|
|
skb_reset_mac_header(skb);
|
|
|
|
if (!IS_ENABLED(CONFIG_IPV6) || family == AF_INET)
|
|
err = IP_ECN_decapsulate(oiph, skb);
|
|
else
|
|
err = IP6_ECN_decapsulate(oiph, skb);
|
|
|
|
if (unlikely(err)) {
|
|
if (log_ecn_error) {
|
|
if (!IS_ENABLED(CONFIG_IPV6) || family == AF_INET)
|
|
net_info_ratelimited("non-ECT from %pI4 "
|
|
"with TOS=%#x\n",
|
|
&((struct iphdr *)oiph)->saddr,
|
|
((struct iphdr *)oiph)->tos);
|
|
else
|
|
net_info_ratelimited("non-ECT from %pI6\n",
|
|
&((struct ipv6hdr *)oiph)->saddr);
|
|
}
|
|
if (err > 1) {
|
|
++bareudp->dev->stats.rx_frame_errors;
|
|
++bareudp->dev->stats.rx_errors;
|
|
goto drop;
|
|
}
|
|
}
|
|
|
|
len = skb->len;
|
|
err = gro_cells_receive(&bareudp->gro_cells, skb);
|
|
if (likely(err == NET_RX_SUCCESS))
|
|
dev_sw_netstats_rx_add(bareudp->dev, len);
|
|
|
|
return 0;
|
|
drop:
|
|
/* Consume bad packet */
|
|
kfree_skb(skb);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int bareudp_err_lookup(struct sock *sk, struct sk_buff *skb)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static int bareudp_init(struct net_device *dev)
|
|
{
|
|
struct bareudp_dev *bareudp = netdev_priv(dev);
|
|
int err;
|
|
|
|
dev->tstats = netdev_alloc_pcpu_stats(struct pcpu_sw_netstats);
|
|
if (!dev->tstats)
|
|
return -ENOMEM;
|
|
|
|
err = gro_cells_init(&bareudp->gro_cells, dev);
|
|
if (err) {
|
|
free_percpu(dev->tstats);
|
|
return err;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static void bareudp_uninit(struct net_device *dev)
|
|
{
|
|
struct bareudp_dev *bareudp = netdev_priv(dev);
|
|
|
|
gro_cells_destroy(&bareudp->gro_cells);
|
|
free_percpu(dev->tstats);
|
|
}
|
|
|
|
static struct socket *bareudp_create_sock(struct net *net, __be16 port)
|
|
{
|
|
struct udp_port_cfg udp_conf;
|
|
struct socket *sock;
|
|
int err;
|
|
|
|
memset(&udp_conf, 0, sizeof(udp_conf));
|
|
#if IS_ENABLED(CONFIG_IPV6)
|
|
udp_conf.family = AF_INET6;
|
|
#else
|
|
udp_conf.family = AF_INET;
|
|
#endif
|
|
udp_conf.local_udp_port = port;
|
|
/* Open UDP socket */
|
|
err = udp_sock_create(net, &udp_conf, &sock);
|
|
if (err < 0)
|
|
return ERR_PTR(err);
|
|
|
|
udp_allow_gso(sock->sk);
|
|
return sock;
|
|
}
|
|
|
|
/* Create new listen socket if needed */
|
|
static int bareudp_socket_create(struct bareudp_dev *bareudp, __be16 port)
|
|
{
|
|
struct udp_tunnel_sock_cfg tunnel_cfg;
|
|
struct socket *sock;
|
|
|
|
sock = bareudp_create_sock(bareudp->net, port);
|
|
if (IS_ERR(sock))
|
|
return PTR_ERR(sock);
|
|
|
|
/* Mark socket as an encapsulation socket */
|
|
memset(&tunnel_cfg, 0, sizeof(tunnel_cfg));
|
|
tunnel_cfg.sk_user_data = bareudp;
|
|
tunnel_cfg.encap_type = 1;
|
|
tunnel_cfg.encap_rcv = bareudp_udp_encap_recv;
|
|
tunnel_cfg.encap_err_lookup = bareudp_err_lookup;
|
|
tunnel_cfg.encap_destroy = NULL;
|
|
setup_udp_tunnel_sock(bareudp->net, sock, &tunnel_cfg);
|
|
|
|
rcu_assign_pointer(bareudp->sock, sock);
|
|
return 0;
|
|
}
|
|
|
|
static int bareudp_open(struct net_device *dev)
|
|
{
|
|
struct bareudp_dev *bareudp = netdev_priv(dev);
|
|
int ret = 0;
|
|
|
|
ret = bareudp_socket_create(bareudp, bareudp->port);
|
|
return ret;
|
|
}
|
|
|
|
static void bareudp_sock_release(struct bareudp_dev *bareudp)
|
|
{
|
|
struct socket *sock;
|
|
|
|
sock = bareudp->sock;
|
|
rcu_assign_pointer(bareudp->sock, NULL);
|
|
synchronize_net();
|
|
udp_tunnel_sock_release(sock);
|
|
}
|
|
|
|
static int bareudp_stop(struct net_device *dev)
|
|
{
|
|
struct bareudp_dev *bareudp = netdev_priv(dev);
|
|
|
|
bareudp_sock_release(bareudp);
|
|
return 0;
|
|
}
|
|
|
|
static int bareudp_xmit_skb(struct sk_buff *skb, struct net_device *dev,
|
|
struct bareudp_dev *bareudp,
|
|
const struct ip_tunnel_info *info)
|
|
{
|
|
bool xnet = !net_eq(bareudp->net, dev_net(bareudp->dev));
|
|
bool use_cache = ip_tunnel_dst_cache_usable(skb, info);
|
|
struct socket *sock = rcu_dereference(bareudp->sock);
|
|
bool udp_sum = !!(info->key.tun_flags & TUNNEL_CSUM);
|
|
const struct ip_tunnel_key *key = &info->key;
|
|
struct rtable *rt;
|
|
__be16 sport, df;
|
|
int min_headroom;
|
|
__u8 tos, ttl;
|
|
__be32 saddr;
|
|
int err;
|
|
|
|
if (!sock)
|
|
return -ESHUTDOWN;
|
|
|
|
rt = ip_route_output_tunnel(skb, dev, bareudp->net, &saddr, info,
|
|
IPPROTO_UDP, use_cache);
|
|
|
|
if (IS_ERR(rt))
|
|
return PTR_ERR(rt);
|
|
|
|
skb_tunnel_check_pmtu(skb, &rt->dst,
|
|
BAREUDP_IPV4_HLEN + info->options_len, false);
|
|
|
|
sport = udp_flow_src_port(bareudp->net, skb,
|
|
bareudp->sport_min, USHRT_MAX,
|
|
true);
|
|
tos = ip_tunnel_ecn_encap(key->tos, ip_hdr(skb), skb);
|
|
ttl = key->ttl;
|
|
df = key->tun_flags & TUNNEL_DONT_FRAGMENT ? htons(IP_DF) : 0;
|
|
skb_scrub_packet(skb, xnet);
|
|
|
|
err = -ENOSPC;
|
|
if (!skb_pull(skb, skb_network_offset(skb)))
|
|
goto free_dst;
|
|
|
|
min_headroom = LL_RESERVED_SPACE(rt->dst.dev) + rt->dst.header_len +
|
|
BAREUDP_BASE_HLEN + info->options_len + sizeof(struct iphdr);
|
|
|
|
err = skb_cow_head(skb, min_headroom);
|
|
if (unlikely(err))
|
|
goto free_dst;
|
|
|
|
err = udp_tunnel_handle_offloads(skb, udp_sum);
|
|
if (err)
|
|
goto free_dst;
|
|
|
|
skb_set_inner_protocol(skb, bareudp->ethertype);
|
|
udp_tunnel_xmit_skb(rt, sock->sk, skb, saddr, info->key.u.ipv4.dst,
|
|
tos, ttl, df, sport, bareudp->port,
|
|
!net_eq(bareudp->net, dev_net(bareudp->dev)),
|
|
!(info->key.tun_flags & TUNNEL_CSUM));
|
|
return 0;
|
|
|
|
free_dst:
|
|
dst_release(&rt->dst);
|
|
return err;
|
|
}
|
|
|
|
static int bareudp6_xmit_skb(struct sk_buff *skb, struct net_device *dev,
|
|
struct bareudp_dev *bareudp,
|
|
const struct ip_tunnel_info *info)
|
|
{
|
|
bool xnet = !net_eq(bareudp->net, dev_net(bareudp->dev));
|
|
bool use_cache = ip_tunnel_dst_cache_usable(skb, info);
|
|
struct socket *sock = rcu_dereference(bareudp->sock);
|
|
bool udp_sum = !!(info->key.tun_flags & TUNNEL_CSUM);
|
|
const struct ip_tunnel_key *key = &info->key;
|
|
struct dst_entry *dst = NULL;
|
|
struct in6_addr saddr, daddr;
|
|
int min_headroom;
|
|
__u8 prio, ttl;
|
|
__be16 sport;
|
|
int err;
|
|
|
|
if (!sock)
|
|
return -ESHUTDOWN;
|
|
|
|
dst = ip6_dst_lookup_tunnel(skb, dev, bareudp->net, sock, &saddr, info,
|
|
IPPROTO_UDP, use_cache);
|
|
if (IS_ERR(dst))
|
|
return PTR_ERR(dst);
|
|
|
|
skb_tunnel_check_pmtu(skb, dst, BAREUDP_IPV6_HLEN + info->options_len,
|
|
false);
|
|
|
|
sport = udp_flow_src_port(bareudp->net, skb,
|
|
bareudp->sport_min, USHRT_MAX,
|
|
true);
|
|
prio = ip_tunnel_ecn_encap(key->tos, ip_hdr(skb), skb);
|
|
ttl = key->ttl;
|
|
|
|
skb_scrub_packet(skb, xnet);
|
|
|
|
err = -ENOSPC;
|
|
if (!skb_pull(skb, skb_network_offset(skb)))
|
|
goto free_dst;
|
|
|
|
min_headroom = LL_RESERVED_SPACE(dst->dev) + dst->header_len +
|
|
BAREUDP_BASE_HLEN + info->options_len + sizeof(struct ipv6hdr);
|
|
|
|
err = skb_cow_head(skb, min_headroom);
|
|
if (unlikely(err))
|
|
goto free_dst;
|
|
|
|
err = udp_tunnel_handle_offloads(skb, udp_sum);
|
|
if (err)
|
|
goto free_dst;
|
|
|
|
daddr = info->key.u.ipv6.dst;
|
|
udp_tunnel6_xmit_skb(dst, sock->sk, skb, dev,
|
|
&saddr, &daddr, prio, ttl,
|
|
info->key.label, sport, bareudp->port,
|
|
!(info->key.tun_flags & TUNNEL_CSUM));
|
|
return 0;
|
|
|
|
free_dst:
|
|
dst_release(dst);
|
|
return err;
|
|
}
|
|
|
|
static bool bareudp_proto_valid(struct bareudp_dev *bareudp, __be16 proto)
|
|
{
|
|
if (bareudp->ethertype == proto)
|
|
return true;
|
|
|
|
if (!bareudp->multi_proto_mode)
|
|
return false;
|
|
|
|
if (bareudp->ethertype == htons(ETH_P_MPLS_UC) &&
|
|
proto == htons(ETH_P_MPLS_MC))
|
|
return true;
|
|
|
|
if (bareudp->ethertype == htons(ETH_P_IP) &&
|
|
proto == htons(ETH_P_IPV6))
|
|
return true;
|
|
|
|
return false;
|
|
}
|
|
|
|
static netdev_tx_t bareudp_xmit(struct sk_buff *skb, struct net_device *dev)
|
|
{
|
|
struct bareudp_dev *bareudp = netdev_priv(dev);
|
|
struct ip_tunnel_info *info = NULL;
|
|
int err;
|
|
|
|
if (!bareudp_proto_valid(bareudp, skb->protocol)) {
|
|
err = -EINVAL;
|
|
goto tx_error;
|
|
}
|
|
|
|
info = skb_tunnel_info(skb);
|
|
if (unlikely(!info || !(info->mode & IP_TUNNEL_INFO_TX))) {
|
|
err = -EINVAL;
|
|
goto tx_error;
|
|
}
|
|
|
|
rcu_read_lock();
|
|
if (IS_ENABLED(CONFIG_IPV6) && info->mode & IP_TUNNEL_INFO_IPV6)
|
|
err = bareudp6_xmit_skb(skb, dev, bareudp, info);
|
|
else
|
|
err = bareudp_xmit_skb(skb, dev, bareudp, info);
|
|
|
|
rcu_read_unlock();
|
|
|
|
if (likely(!err))
|
|
return NETDEV_TX_OK;
|
|
tx_error:
|
|
dev_kfree_skb(skb);
|
|
|
|
if (err == -ELOOP)
|
|
dev->stats.collisions++;
|
|
else if (err == -ENETUNREACH)
|
|
dev->stats.tx_carrier_errors++;
|
|
|
|
dev->stats.tx_errors++;
|
|
return NETDEV_TX_OK;
|
|
}
|
|
|
|
static int bareudp_fill_metadata_dst(struct net_device *dev,
|
|
struct sk_buff *skb)
|
|
{
|
|
struct ip_tunnel_info *info = skb_tunnel_info(skb);
|
|
struct bareudp_dev *bareudp = netdev_priv(dev);
|
|
bool use_cache;
|
|
|
|
use_cache = ip_tunnel_dst_cache_usable(skb, info);
|
|
|
|
if (!IS_ENABLED(CONFIG_IPV6) || ip_tunnel_info_af(info) == AF_INET) {
|
|
struct rtable *rt;
|
|
__be32 saddr;
|
|
|
|
rt = ip_route_output_tunnel(skb, dev, bareudp->net, &saddr,
|
|
info, IPPROTO_UDP, use_cache);
|
|
if (IS_ERR(rt))
|
|
return PTR_ERR(rt);
|
|
|
|
ip_rt_put(rt);
|
|
info->key.u.ipv4.src = saddr;
|
|
} else if (ip_tunnel_info_af(info) == AF_INET6) {
|
|
struct dst_entry *dst;
|
|
struct in6_addr saddr;
|
|
struct socket *sock = rcu_dereference(bareudp->sock);
|
|
|
|
dst = ip6_dst_lookup_tunnel(skb, dev, bareudp->net, sock,
|
|
&saddr, info, IPPROTO_UDP,
|
|
use_cache);
|
|
if (IS_ERR(dst))
|
|
return PTR_ERR(dst);
|
|
|
|
dst_release(dst);
|
|
info->key.u.ipv6.src = saddr;
|
|
} else {
|
|
return -EINVAL;
|
|
}
|
|
|
|
info->key.tp_src = udp_flow_src_port(bareudp->net, skb,
|
|
bareudp->sport_min,
|
|
USHRT_MAX, true);
|
|
info->key.tp_dst = bareudp->port;
|
|
return 0;
|
|
}
|
|
|
|
static const struct net_device_ops bareudp_netdev_ops = {
|
|
.ndo_init = bareudp_init,
|
|
.ndo_uninit = bareudp_uninit,
|
|
.ndo_open = bareudp_open,
|
|
.ndo_stop = bareudp_stop,
|
|
.ndo_start_xmit = bareudp_xmit,
|
|
.ndo_get_stats64 = dev_get_tstats64,
|
|
.ndo_fill_metadata_dst = bareudp_fill_metadata_dst,
|
|
};
|
|
|
|
static const struct nla_policy bareudp_policy[IFLA_BAREUDP_MAX + 1] = {
|
|
[IFLA_BAREUDP_PORT] = { .type = NLA_U16 },
|
|
[IFLA_BAREUDP_ETHERTYPE] = { .type = NLA_U16 },
|
|
[IFLA_BAREUDP_SRCPORT_MIN] = { .type = NLA_U16 },
|
|
[IFLA_BAREUDP_MULTIPROTO_MODE] = { .type = NLA_FLAG },
|
|
};
|
|
|
|
/* Info for udev, that this is a virtual tunnel endpoint */
|
|
static const struct device_type bareudp_type = {
|
|
.name = "bareudp",
|
|
};
|
|
|
|
/* Initialize the device structure. */
|
|
static void bareudp_setup(struct net_device *dev)
|
|
{
|
|
dev->netdev_ops = &bareudp_netdev_ops;
|
|
dev->needs_free_netdev = true;
|
|
SET_NETDEV_DEVTYPE(dev, &bareudp_type);
|
|
dev->features |= NETIF_F_SG | NETIF_F_HW_CSUM | NETIF_F_FRAGLIST;
|
|
dev->features |= NETIF_F_RXCSUM;
|
|
dev->features |= NETIF_F_LLTX;
|
|
dev->features |= NETIF_F_GSO_SOFTWARE;
|
|
dev->hw_features |= NETIF_F_SG | NETIF_F_HW_CSUM | NETIF_F_FRAGLIST;
|
|
dev->hw_features |= NETIF_F_RXCSUM;
|
|
dev->hw_features |= NETIF_F_GSO_SOFTWARE;
|
|
dev->hard_header_len = 0;
|
|
dev->addr_len = 0;
|
|
dev->mtu = ETH_DATA_LEN;
|
|
dev->min_mtu = IPV4_MIN_MTU;
|
|
dev->max_mtu = IP_MAX_MTU - BAREUDP_BASE_HLEN;
|
|
dev->type = ARPHRD_NONE;
|
|
netif_keep_dst(dev);
|
|
dev->priv_flags |= IFF_NO_QUEUE;
|
|
dev->flags = IFF_POINTOPOINT | IFF_NOARP | IFF_MULTICAST;
|
|
}
|
|
|
|
static int bareudp_validate(struct nlattr *tb[], struct nlattr *data[],
|
|
struct netlink_ext_ack *extack)
|
|
{
|
|
if (!data) {
|
|
NL_SET_ERR_MSG(extack,
|
|
"Not enough attributes provided to perform the operation");
|
|
return -EINVAL;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int bareudp2info(struct nlattr *data[], struct bareudp_conf *conf,
|
|
struct netlink_ext_ack *extack)
|
|
{
|
|
memset(conf, 0, sizeof(*conf));
|
|
|
|
if (!data[IFLA_BAREUDP_PORT]) {
|
|
NL_SET_ERR_MSG(extack, "port not specified");
|
|
return -EINVAL;
|
|
}
|
|
if (!data[IFLA_BAREUDP_ETHERTYPE]) {
|
|
NL_SET_ERR_MSG(extack, "ethertype not specified");
|
|
return -EINVAL;
|
|
}
|
|
|
|
conf->port = nla_get_u16(data[IFLA_BAREUDP_PORT]);
|
|
conf->ethertype = nla_get_u16(data[IFLA_BAREUDP_ETHERTYPE]);
|
|
|
|
if (data[IFLA_BAREUDP_SRCPORT_MIN])
|
|
conf->sport_min = nla_get_u16(data[IFLA_BAREUDP_SRCPORT_MIN]);
|
|
|
|
if (data[IFLA_BAREUDP_MULTIPROTO_MODE])
|
|
conf->multi_proto_mode = true;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static struct bareudp_dev *bareudp_find_dev(struct bareudp_net *bn,
|
|
const struct bareudp_conf *conf)
|
|
{
|
|
struct bareudp_dev *bareudp, *t = NULL;
|
|
|
|
list_for_each_entry(bareudp, &bn->bareudp_list, next) {
|
|
if (conf->port == bareudp->port)
|
|
t = bareudp;
|
|
}
|
|
return t;
|
|
}
|
|
|
|
static int bareudp_configure(struct net *net, struct net_device *dev,
|
|
struct bareudp_conf *conf,
|
|
struct netlink_ext_ack *extack)
|
|
{
|
|
struct bareudp_net *bn = net_generic(net, bareudp_net_id);
|
|
struct bareudp_dev *t, *bareudp = netdev_priv(dev);
|
|
int err;
|
|
|
|
bareudp->net = net;
|
|
bareudp->dev = dev;
|
|
t = bareudp_find_dev(bn, conf);
|
|
if (t) {
|
|
NL_SET_ERR_MSG(extack, "Another bareudp device using the same port already exists");
|
|
return -EBUSY;
|
|
}
|
|
|
|
if (conf->multi_proto_mode &&
|
|
(conf->ethertype != htons(ETH_P_MPLS_UC) &&
|
|
conf->ethertype != htons(ETH_P_IP))) {
|
|
NL_SET_ERR_MSG(extack, "Cannot set multiproto mode for this ethertype (only IPv4 and unicast MPLS are supported)");
|
|
return -EINVAL;
|
|
}
|
|
|
|
bareudp->port = conf->port;
|
|
bareudp->ethertype = conf->ethertype;
|
|
bareudp->sport_min = conf->sport_min;
|
|
bareudp->multi_proto_mode = conf->multi_proto_mode;
|
|
|
|
err = register_netdevice(dev);
|
|
if (err)
|
|
return err;
|
|
|
|
list_add(&bareudp->next, &bn->bareudp_list);
|
|
return 0;
|
|
}
|
|
|
|
static int bareudp_link_config(struct net_device *dev,
|
|
struct nlattr *tb[])
|
|
{
|
|
int err;
|
|
|
|
if (tb[IFLA_MTU]) {
|
|
err = dev_set_mtu(dev, nla_get_u32(tb[IFLA_MTU]));
|
|
if (err)
|
|
return err;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static void bareudp_dellink(struct net_device *dev, struct list_head *head)
|
|
{
|
|
struct bareudp_dev *bareudp = netdev_priv(dev);
|
|
|
|
list_del(&bareudp->next);
|
|
unregister_netdevice_queue(dev, head);
|
|
}
|
|
|
|
static int bareudp_newlink(struct net *net, struct net_device *dev,
|
|
struct nlattr *tb[], struct nlattr *data[],
|
|
struct netlink_ext_ack *extack)
|
|
{
|
|
struct bareudp_conf conf;
|
|
int err;
|
|
|
|
err = bareudp2info(data, &conf, extack);
|
|
if (err)
|
|
return err;
|
|
|
|
err = bareudp_configure(net, dev, &conf, extack);
|
|
if (err)
|
|
return err;
|
|
|
|
err = bareudp_link_config(dev, tb);
|
|
if (err)
|
|
goto err_unconfig;
|
|
|
|
return 0;
|
|
|
|
err_unconfig:
|
|
bareudp_dellink(dev, NULL);
|
|
return err;
|
|
}
|
|
|
|
static size_t bareudp_get_size(const struct net_device *dev)
|
|
{
|
|
return nla_total_size(sizeof(__be16)) + /* IFLA_BAREUDP_PORT */
|
|
nla_total_size(sizeof(__be16)) + /* IFLA_BAREUDP_ETHERTYPE */
|
|
nla_total_size(sizeof(__u16)) + /* IFLA_BAREUDP_SRCPORT_MIN */
|
|
nla_total_size(0) + /* IFLA_BAREUDP_MULTIPROTO_MODE */
|
|
0;
|
|
}
|
|
|
|
static int bareudp_fill_info(struct sk_buff *skb, const struct net_device *dev)
|
|
{
|
|
struct bareudp_dev *bareudp = netdev_priv(dev);
|
|
|
|
if (nla_put_be16(skb, IFLA_BAREUDP_PORT, bareudp->port))
|
|
goto nla_put_failure;
|
|
if (nla_put_be16(skb, IFLA_BAREUDP_ETHERTYPE, bareudp->ethertype))
|
|
goto nla_put_failure;
|
|
if (nla_put_u16(skb, IFLA_BAREUDP_SRCPORT_MIN, bareudp->sport_min))
|
|
goto nla_put_failure;
|
|
if (bareudp->multi_proto_mode &&
|
|
nla_put_flag(skb, IFLA_BAREUDP_MULTIPROTO_MODE))
|
|
goto nla_put_failure;
|
|
|
|
return 0;
|
|
|
|
nla_put_failure:
|
|
return -EMSGSIZE;
|
|
}
|
|
|
|
static struct rtnl_link_ops bareudp_link_ops __read_mostly = {
|
|
.kind = "bareudp",
|
|
.maxtype = IFLA_BAREUDP_MAX,
|
|
.policy = bareudp_policy,
|
|
.priv_size = sizeof(struct bareudp_dev),
|
|
.setup = bareudp_setup,
|
|
.validate = bareudp_validate,
|
|
.newlink = bareudp_newlink,
|
|
.dellink = bareudp_dellink,
|
|
.get_size = bareudp_get_size,
|
|
.fill_info = bareudp_fill_info,
|
|
};
|
|
|
|
static __net_init int bareudp_init_net(struct net *net)
|
|
{
|
|
struct bareudp_net *bn = net_generic(net, bareudp_net_id);
|
|
|
|
INIT_LIST_HEAD(&bn->bareudp_list);
|
|
return 0;
|
|
}
|
|
|
|
static void bareudp_destroy_tunnels(struct net *net, struct list_head *head)
|
|
{
|
|
struct bareudp_net *bn = net_generic(net, bareudp_net_id);
|
|
struct bareudp_dev *bareudp, *next;
|
|
|
|
list_for_each_entry_safe(bareudp, next, &bn->bareudp_list, next)
|
|
unregister_netdevice_queue(bareudp->dev, head);
|
|
}
|
|
|
|
static void __net_exit bareudp_exit_batch_net(struct list_head *net_list)
|
|
{
|
|
struct net *net;
|
|
LIST_HEAD(list);
|
|
|
|
rtnl_lock();
|
|
list_for_each_entry(net, net_list, exit_list)
|
|
bareudp_destroy_tunnels(net, &list);
|
|
|
|
/* unregister the devices gathered above */
|
|
unregister_netdevice_many(&list);
|
|
rtnl_unlock();
|
|
}
|
|
|
|
static struct pernet_operations bareudp_net_ops = {
|
|
.init = bareudp_init_net,
|
|
.exit_batch = bareudp_exit_batch_net,
|
|
.id = &bareudp_net_id,
|
|
.size = sizeof(struct bareudp_net),
|
|
};
|
|
|
|
static int __init bareudp_init_module(void)
|
|
{
|
|
int rc;
|
|
|
|
rc = register_pernet_subsys(&bareudp_net_ops);
|
|
if (rc)
|
|
goto out1;
|
|
|
|
rc = rtnl_link_register(&bareudp_link_ops);
|
|
if (rc)
|
|
goto out2;
|
|
|
|
return 0;
|
|
out2:
|
|
unregister_pernet_subsys(&bareudp_net_ops);
|
|
out1:
|
|
return rc;
|
|
}
|
|
late_initcall(bareudp_init_module);
|
|
|
|
static void __exit bareudp_cleanup_module(void)
|
|
{
|
|
rtnl_link_unregister(&bareudp_link_ops);
|
|
unregister_pernet_subsys(&bareudp_net_ops);
|
|
}
|
|
module_exit(bareudp_cleanup_module);
|
|
|
|
MODULE_ALIAS_RTNL_LINK("bareudp");
|
|
MODULE_LICENSE("GPL");
|
|
MODULE_AUTHOR("Martin Varghese <martin.varghese@nokia.com>");
|
|
MODULE_DESCRIPTION("Interface driver for UDP encapsulated traffic");
|