rmnet_vnd.c (8783B)
1// SPDX-License-Identifier: GPL-2.0-only 2/* Copyright (c) 2013-2018, The Linux Foundation. All rights reserved. 3 * 4 * RMNET Data virtual network driver 5 */ 6 7#include <linux/etherdevice.h> 8#include <linux/ethtool.h> 9#include <linux/if_arp.h> 10#include <net/pkt_sched.h> 11#include "rmnet_config.h" 12#include "rmnet_handlers.h" 13#include "rmnet_private.h" 14#include "rmnet_map.h" 15#include "rmnet_vnd.h" 16 17/* RX/TX Fixup */ 18 19void rmnet_vnd_rx_fixup(struct sk_buff *skb, struct net_device *dev) 20{ 21 struct rmnet_priv *priv = netdev_priv(dev); 22 struct rmnet_pcpu_stats *pcpu_ptr; 23 24 pcpu_ptr = this_cpu_ptr(priv->pcpu_stats); 25 26 u64_stats_update_begin(&pcpu_ptr->syncp); 27 pcpu_ptr->stats.rx_pkts++; 28 pcpu_ptr->stats.rx_bytes += skb->len; 29 u64_stats_update_end(&pcpu_ptr->syncp); 30} 31 32void rmnet_vnd_tx_fixup(struct sk_buff *skb, struct net_device *dev) 33{ 34 struct rmnet_priv *priv = netdev_priv(dev); 35 struct rmnet_pcpu_stats *pcpu_ptr; 36 37 pcpu_ptr = this_cpu_ptr(priv->pcpu_stats); 38 39 u64_stats_update_begin(&pcpu_ptr->syncp); 40 pcpu_ptr->stats.tx_pkts++; 41 pcpu_ptr->stats.tx_bytes += skb->len; 42 u64_stats_update_end(&pcpu_ptr->syncp); 43} 44 45/* Network Device Operations */ 46 47static netdev_tx_t rmnet_vnd_start_xmit(struct sk_buff *skb, 48 struct net_device *dev) 49{ 50 struct rmnet_priv *priv; 51 52 priv = netdev_priv(dev); 53 if (priv->real_dev) { 54 rmnet_egress_handler(skb); 55 } else { 56 this_cpu_inc(priv->pcpu_stats->stats.tx_drops); 57 kfree_skb(skb); 58 } 59 return NETDEV_TX_OK; 60} 61 62static int rmnet_vnd_headroom(struct rmnet_port *port) 63{ 64 u32 headroom; 65 66 headroom = sizeof(struct rmnet_map_header); 67 68 if (port->data_format & RMNET_FLAGS_EGRESS_MAP_CKSUMV4) 69 headroom += sizeof(struct rmnet_map_ul_csum_header); 70 71 return headroom; 72} 73 74static int rmnet_vnd_change_mtu(struct net_device *rmnet_dev, int new_mtu) 75{ 76 struct rmnet_priv *priv = netdev_priv(rmnet_dev); 77 struct rmnet_port *port; 78 u32 headroom; 79 80 port = rmnet_get_port_rtnl(priv->real_dev); 81 82 headroom = rmnet_vnd_headroom(port); 83 84 if (new_mtu < 0 || new_mtu > RMNET_MAX_PACKET_SIZE || 85 new_mtu > (priv->real_dev->mtu - headroom)) 86 return -EINVAL; 87 88 rmnet_dev->mtu = new_mtu; 89 return 0; 90} 91 92static int rmnet_vnd_get_iflink(const struct net_device *dev) 93{ 94 struct rmnet_priv *priv = netdev_priv(dev); 95 96 return priv->real_dev->ifindex; 97} 98 99static int rmnet_vnd_init(struct net_device *dev) 100{ 101 struct rmnet_priv *priv = netdev_priv(dev); 102 int err; 103 104 priv->pcpu_stats = alloc_percpu(struct rmnet_pcpu_stats); 105 if (!priv->pcpu_stats) 106 return -ENOMEM; 107 108 err = gro_cells_init(&priv->gro_cells, dev); 109 if (err) { 110 free_percpu(priv->pcpu_stats); 111 return err; 112 } 113 114 return 0; 115} 116 117static void rmnet_vnd_uninit(struct net_device *dev) 118{ 119 struct rmnet_priv *priv = netdev_priv(dev); 120 121 gro_cells_destroy(&priv->gro_cells); 122 free_percpu(priv->pcpu_stats); 123} 124 125static void rmnet_get_stats64(struct net_device *dev, 126 struct rtnl_link_stats64 *s) 127{ 128 struct rmnet_priv *priv = netdev_priv(dev); 129 struct rmnet_vnd_stats total_stats = { }; 130 struct rmnet_pcpu_stats *pcpu_ptr; 131 struct rmnet_vnd_stats snapshot; 132 unsigned int cpu, start; 133 134 for_each_possible_cpu(cpu) { 135 pcpu_ptr = per_cpu_ptr(priv->pcpu_stats, cpu); 136 137 do { 138 start = u64_stats_fetch_begin_irq(&pcpu_ptr->syncp); 139 snapshot = pcpu_ptr->stats; /* struct assignment */ 140 } while (u64_stats_fetch_retry_irq(&pcpu_ptr->syncp, start)); 141 142 total_stats.rx_pkts += snapshot.rx_pkts; 143 total_stats.rx_bytes += snapshot.rx_bytes; 144 total_stats.tx_pkts += snapshot.tx_pkts; 145 total_stats.tx_bytes += snapshot.tx_bytes; 146 total_stats.tx_drops += snapshot.tx_drops; 147 } 148 149 s->rx_packets = total_stats.rx_pkts; 150 s->rx_bytes = total_stats.rx_bytes; 151 s->tx_packets = total_stats.tx_pkts; 152 s->tx_bytes = total_stats.tx_bytes; 153 s->tx_dropped = total_stats.tx_drops; 154} 155 156static const struct net_device_ops rmnet_vnd_ops = { 157 .ndo_start_xmit = rmnet_vnd_start_xmit, 158 .ndo_change_mtu = rmnet_vnd_change_mtu, 159 .ndo_get_iflink = rmnet_vnd_get_iflink, 160 .ndo_add_slave = rmnet_add_bridge, 161 .ndo_del_slave = rmnet_del_bridge, 162 .ndo_init = rmnet_vnd_init, 163 .ndo_uninit = rmnet_vnd_uninit, 164 .ndo_get_stats64 = rmnet_get_stats64, 165}; 166 167static const char rmnet_gstrings_stats[][ETH_GSTRING_LEN] = { 168 "Checksum ok", 169 "Bad IPv4 header checksum", 170 "Checksum valid bit not set", 171 "Checksum validation failed", 172 "Checksum error bad buffer", 173 "Checksum error bad ip version", 174 "Checksum error bad transport", 175 "Checksum skipped on ip fragment", 176 "Checksum skipped", 177 "Checksum computed in software", 178 "Checksum computed in hardware", 179}; 180 181static void rmnet_get_strings(struct net_device *dev, u32 stringset, u8 *buf) 182{ 183 switch (stringset) { 184 case ETH_SS_STATS: 185 memcpy(buf, &rmnet_gstrings_stats, 186 sizeof(rmnet_gstrings_stats)); 187 break; 188 } 189} 190 191static int rmnet_get_sset_count(struct net_device *dev, int sset) 192{ 193 switch (sset) { 194 case ETH_SS_STATS: 195 return ARRAY_SIZE(rmnet_gstrings_stats); 196 default: 197 return -EOPNOTSUPP; 198 } 199} 200 201static void rmnet_get_ethtool_stats(struct net_device *dev, 202 struct ethtool_stats *stats, u64 *data) 203{ 204 struct rmnet_priv *priv = netdev_priv(dev); 205 struct rmnet_priv_stats *st = &priv->stats; 206 207 if (!data) 208 return; 209 210 memcpy(data, st, ARRAY_SIZE(rmnet_gstrings_stats) * sizeof(u64)); 211} 212 213static const struct ethtool_ops rmnet_ethtool_ops = { 214 .get_ethtool_stats = rmnet_get_ethtool_stats, 215 .get_strings = rmnet_get_strings, 216 .get_sset_count = rmnet_get_sset_count, 217}; 218 219/* Called by kernel whenever a new rmnet<n> device is created. Sets MTU, 220 * flags, ARP type, needed headroom, etc... 221 */ 222void rmnet_vnd_setup(struct net_device *rmnet_dev) 223{ 224 rmnet_dev->netdev_ops = &rmnet_vnd_ops; 225 rmnet_dev->mtu = RMNET_DFLT_PACKET_SIZE; 226 rmnet_dev->needed_headroom = RMNET_NEEDED_HEADROOM; 227 eth_hw_addr_random(rmnet_dev); 228 rmnet_dev->tx_queue_len = RMNET_TX_QUEUE_LEN; 229 230 /* Raw IP mode */ 231 rmnet_dev->header_ops = NULL; /* No header */ 232 rmnet_dev->type = ARPHRD_RAWIP; 233 rmnet_dev->hard_header_len = 0; 234 rmnet_dev->flags &= ~(IFF_BROADCAST | IFF_MULTICAST); 235 236 rmnet_dev->needs_free_netdev = true; 237 rmnet_dev->ethtool_ops = &rmnet_ethtool_ops; 238 239 rmnet_dev->features |= NETIF_F_LLTX; 240 241 /* This perm addr will be used as interface identifier by IPv6 */ 242 rmnet_dev->addr_assign_type = NET_ADDR_RANDOM; 243 eth_random_addr(rmnet_dev->perm_addr); 244} 245 246/* Exposed API */ 247 248int rmnet_vnd_newlink(u8 id, struct net_device *rmnet_dev, 249 struct rmnet_port *port, 250 struct net_device *real_dev, 251 struct rmnet_endpoint *ep, 252 struct netlink_ext_ack *extack) 253 254{ 255 struct rmnet_priv *priv = netdev_priv(rmnet_dev); 256 u32 headroom; 257 int rc; 258 259 if (rmnet_get_endpoint(port, id)) { 260 NL_SET_ERR_MSG_MOD(extack, "MUX ID already exists"); 261 return -EBUSY; 262 } 263 264 rmnet_dev->hw_features = NETIF_F_RXCSUM; 265 rmnet_dev->hw_features |= NETIF_F_IP_CSUM | NETIF_F_IPV6_CSUM; 266 rmnet_dev->hw_features |= NETIF_F_SG; 267 268 priv->real_dev = real_dev; 269 270 headroom = rmnet_vnd_headroom(port); 271 272 if (rmnet_vnd_change_mtu(rmnet_dev, real_dev->mtu - headroom)) { 273 NL_SET_ERR_MSG_MOD(extack, "Invalid MTU on real dev"); 274 return -EINVAL; 275 } 276 277 rc = register_netdevice(rmnet_dev); 278 if (!rc) { 279 ep->egress_dev = rmnet_dev; 280 ep->mux_id = id; 281 port->nr_rmnet_devs++; 282 283 rmnet_dev->rtnl_link_ops = &rmnet_link_ops; 284 285 priv->mux_id = id; 286 287 netdev_dbg(rmnet_dev, "rmnet dev created\n"); 288 } 289 290 return rc; 291} 292 293int rmnet_vnd_dellink(u8 id, struct rmnet_port *port, 294 struct rmnet_endpoint *ep) 295{ 296 if (id >= RMNET_MAX_LOGICAL_EP || !ep->egress_dev) 297 return -EINVAL; 298 299 ep->egress_dev = NULL; 300 port->nr_rmnet_devs--; 301 return 0; 302} 303 304int rmnet_vnd_do_flow_control(struct net_device *rmnet_dev, int enable) 305{ 306 netdev_dbg(rmnet_dev, "Setting VND TX queue state to %d\n", enable); 307 /* Although we expect similar number of enable/disable 308 * commands, optimize for the disable. That is more 309 * latency sensitive than enable 310 */ 311 if (unlikely(enable)) 312 netif_wake_queue(rmnet_dev); 313 else 314 netif_stop_queue(rmnet_dev); 315 316 return 0; 317} 318 319int rmnet_vnd_validate_real_dev_mtu(struct net_device *real_dev) 320{ 321 struct hlist_node *tmp_ep; 322 struct rmnet_endpoint *ep; 323 struct rmnet_port *port; 324 unsigned long bkt_ep; 325 u32 headroom; 326 327 port = rmnet_get_port_rtnl(real_dev); 328 329 headroom = rmnet_vnd_headroom(port); 330 331 hash_for_each_safe(port->muxed_ep, bkt_ep, tmp_ep, ep, hlnode) { 332 if (ep->egress_dev->mtu > (real_dev->mtu - headroom)) 333 return -1; 334 } 335 336 return 0; 337} 338 339int rmnet_vnd_update_dev_mtu(struct rmnet_port *port, 340 struct net_device *real_dev) 341{ 342 struct hlist_node *tmp_ep; 343 struct rmnet_endpoint *ep; 344 unsigned long bkt_ep; 345 u32 headroom; 346 347 headroom = rmnet_vnd_headroom(port); 348 349 hash_for_each_safe(port->muxed_ep, bkt_ep, tmp_ep, ep, hlnode) { 350 if (ep->egress_dev->mtu <= (real_dev->mtu - headroom)) 351 continue; 352 353 if (rmnet_vnd_change_mtu(ep->egress_dev, 354 real_dev->mtu - headroom)) 355 return -1; 356 } 357 358 return 0; 359}