123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138 |
- // SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB
- /*
- * Copyright (c) 2020 Mellanox Technologies. All rights reserved.
- */
- #include <rdma/ib_verbs.h>
- #include <rdma/ib_cache.h>
- #include <rdma/lag.h>
- static struct sk_buff *rdma_build_skb(struct ib_device *device,
- struct net_device *netdev,
- struct rdma_ah_attr *ah_attr,
- gfp_t flags)
- {
- struct ipv6hdr *ip6h;
- struct sk_buff *skb;
- struct ethhdr *eth;
- struct iphdr *iph;
- struct udphdr *uh;
- u8 smac[ETH_ALEN];
- bool is_ipv4;
- int hdr_len;
- is_ipv4 = ipv6_addr_v4mapped((struct in6_addr *)ah_attr->grh.dgid.raw);
- hdr_len = ETH_HLEN + sizeof(struct udphdr) + LL_RESERVED_SPACE(netdev);
- hdr_len += is_ipv4 ? sizeof(struct iphdr) : sizeof(struct ipv6hdr);
- skb = alloc_skb(hdr_len, flags);
- if (!skb)
- return NULL;
- skb->dev = netdev;
- skb_reserve(skb, hdr_len);
- skb_push(skb, sizeof(struct udphdr));
- skb_reset_transport_header(skb);
- uh = udp_hdr(skb);
- uh->source =
- htons(rdma_flow_label_to_udp_sport(ah_attr->grh.flow_label));
- uh->dest = htons(ROCE_V2_UDP_DPORT);
- uh->len = htons(sizeof(struct udphdr));
- if (is_ipv4) {
- skb_push(skb, sizeof(struct iphdr));
- skb_reset_network_header(skb);
- iph = ip_hdr(skb);
- iph->frag_off = 0;
- iph->version = 4;
- iph->protocol = IPPROTO_UDP;
- iph->ihl = 0x5;
- iph->tot_len = htons(sizeof(struct udphdr) + sizeof(struct
- iphdr));
- memcpy(&iph->saddr, ah_attr->grh.sgid_attr->gid.raw + 12,
- sizeof(struct in_addr));
- memcpy(&iph->daddr, ah_attr->grh.dgid.raw + 12,
- sizeof(struct in_addr));
- } else {
- skb_push(skb, sizeof(struct ipv6hdr));
- skb_reset_network_header(skb);
- ip6h = ipv6_hdr(skb);
- ip6h->version = 6;
- ip6h->nexthdr = IPPROTO_UDP;
- memcpy(&ip6h->flow_lbl, &ah_attr->grh.flow_label,
- sizeof(*ip6h->flow_lbl));
- memcpy(&ip6h->saddr, ah_attr->grh.sgid_attr->gid.raw,
- sizeof(struct in6_addr));
- memcpy(&ip6h->daddr, ah_attr->grh.dgid.raw,
- sizeof(struct in6_addr));
- }
- skb_push(skb, sizeof(struct ethhdr));
- skb_reset_mac_header(skb);
- eth = eth_hdr(skb);
- skb->protocol = eth->h_proto = htons(is_ipv4 ? ETH_P_IP : ETH_P_IPV6);
- rdma_read_gid_l2_fields(ah_attr->grh.sgid_attr, NULL, smac);
- memcpy(eth->h_source, smac, ETH_ALEN);
- memcpy(eth->h_dest, ah_attr->roce.dmac, ETH_ALEN);
- return skb;
- }
- static struct net_device *rdma_get_xmit_slave_udp(struct ib_device *device,
- struct net_device *master,
- struct rdma_ah_attr *ah_attr,
- gfp_t flags)
- {
- struct net_device *slave;
- struct sk_buff *skb;
- skb = rdma_build_skb(device, master, ah_attr, flags);
- if (!skb)
- return ERR_PTR(-ENOMEM);
- rcu_read_lock();
- slave = netdev_get_xmit_slave(master, skb,
- !!(device->lag_flags &
- RDMA_LAG_FLAGS_HASH_ALL_SLAVES));
- if (slave)
- dev_hold(slave);
- rcu_read_unlock();
- kfree_skb(skb);
- return slave;
- }
- void rdma_lag_put_ah_roce_slave(struct net_device *xmit_slave)
- {
- if (xmit_slave)
- dev_put(xmit_slave);
- }
- struct net_device *rdma_lag_get_ah_roce_slave(struct ib_device *device,
- struct rdma_ah_attr *ah_attr,
- gfp_t flags)
- {
- struct net_device *slave = NULL;
- struct net_device *master;
- if (!(ah_attr->type == RDMA_AH_ATTR_TYPE_ROCE &&
- ah_attr->grh.sgid_attr->gid_type == IB_GID_TYPE_ROCE_UDP_ENCAP &&
- ah_attr->grh.flow_label))
- return NULL;
- rcu_read_lock();
- master = rdma_read_gid_attr_ndev_rcu(ah_attr->grh.sgid_attr);
- if (IS_ERR(master)) {
- rcu_read_unlock();
- return master;
- }
- dev_hold(master);
- rcu_read_unlock();
- if (!netif_is_bond_master(master))
- goto put;
- slave = rdma_get_xmit_slave_udp(device, master, ah_attr, flags);
- put:
- dev_put(master);
- return slave;
- }
|