| /* |
| * Copyright (c) 2016, Mellanox Technologies. All rights reserved. |
| * |
| * This software is available to you under a choice of one of two |
| * licenses. You may choose to be licensed under the terms of the GNU |
| * General Public License (GPL) Version 2, available from the file |
| * COPYING in the main directory of this source tree, or the |
| * OpenIB.org BSD license below: |
| * |
| * Redistribution and use in source and binary forms, with or |
| * without modification, are permitted provided that the following |
| * conditions are met: |
| * |
| * - Redistributions of source code must retain the above |
| * copyright notice, this list of conditions and the following |
| * disclaimer. |
| * |
| * - Redistributions in binary form must reproduce the above |
| * copyright notice, this list of conditions and the following |
| * disclaimer in the documentation and/or other materials |
| * provided with the distribution. |
| * |
| * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, |
| * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF |
| * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND |
| * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS |
| * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN |
| * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN |
| * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE |
| * SOFTWARE. |
| */ |
| |
| #include <linux/mlx5/fs.h> |
| #include "en.h" |
| #include "en/params.h" |
| #include "en/xsk/pool.h" |
| |
| struct mlx5e_ethtool_rule { |
| struct list_head list; |
| struct ethtool_rx_flow_spec flow_spec; |
| struct mlx5_flow_handle *rule; |
| struct mlx5e_ethtool_table *eth_ft; |
| }; |
| |
| static void put_flow_table(struct mlx5e_ethtool_table *eth_ft) |
| { |
| if (!--eth_ft->num_rules) { |
| mlx5_destroy_flow_table(eth_ft->ft); |
| eth_ft->ft = NULL; |
| } |
| } |
| |
| #define MLX5E_ETHTOOL_L3_L4_PRIO 0 |
| #define MLX5E_ETHTOOL_L2_PRIO (MLX5E_ETHTOOL_L3_L4_PRIO + ETHTOOL_NUM_L3_L4_FTS) |
| #define MLX5E_ETHTOOL_NUM_ENTRIES 64000 |
| #define MLX5E_ETHTOOL_NUM_GROUPS 10 |
| static struct mlx5e_ethtool_table *get_flow_table(struct mlx5e_priv *priv, |
| struct ethtool_rx_flow_spec *fs, |
| int num_tuples) |
| { |
| struct mlx5_flow_table_attr ft_attr = {}; |
| struct mlx5e_ethtool_table *eth_ft; |
| struct mlx5_flow_namespace *ns; |
| struct mlx5_flow_table *ft; |
| int max_tuples; |
| int table_size; |
| int prio; |
| |
| switch (fs->flow_type & ~(FLOW_EXT | FLOW_MAC_EXT)) { |
| case TCP_V4_FLOW: |
| case UDP_V4_FLOW: |
| case TCP_V6_FLOW: |
| case UDP_V6_FLOW: |
| max_tuples = ETHTOOL_NUM_L3_L4_FTS; |
| prio = MLX5E_ETHTOOL_L3_L4_PRIO + (max_tuples - num_tuples); |
| eth_ft = &priv->fs.ethtool.l3_l4_ft[prio]; |
| break; |
| case IP_USER_FLOW: |
| case IPV6_USER_FLOW: |
| max_tuples = ETHTOOL_NUM_L3_L4_FTS; |
| prio = MLX5E_ETHTOOL_L3_L4_PRIO + (max_tuples - num_tuples); |
| eth_ft = &priv->fs.ethtool.l3_l4_ft[prio]; |
| break; |
| case ETHER_FLOW: |
| max_tuples = ETHTOOL_NUM_L2_FTS; |
| prio = max_tuples - num_tuples; |
| eth_ft = &priv->fs.ethtool.l2_ft[prio]; |
| prio += MLX5E_ETHTOOL_L2_PRIO; |
| break; |
| default: |
| return ERR_PTR(-EINVAL); |
| } |
| |
| eth_ft->num_rules++; |
| if (eth_ft->ft) |
| return eth_ft; |
| |
| ns = mlx5_get_flow_namespace(priv->mdev, |
| MLX5_FLOW_NAMESPACE_ETHTOOL); |
| if (!ns) |
| return ERR_PTR(-EOPNOTSUPP); |
| |
| table_size = min_t(u32, BIT(MLX5_CAP_FLOWTABLE(priv->mdev, |
| flow_table_properties_nic_receive.log_max_ft_size)), |
| MLX5E_ETHTOOL_NUM_ENTRIES); |
| |
| ft_attr.prio = prio; |
| ft_attr.max_fte = table_size; |
| ft_attr.autogroup.max_num_groups = MLX5E_ETHTOOL_NUM_GROUPS; |
| ft = mlx5_create_auto_grouped_flow_table(ns, &ft_attr); |
| if (IS_ERR(ft)) |
| return (void *)ft; |
| |
| eth_ft->ft = ft; |
| return eth_ft; |
| } |
| |
| static void mask_spec(u8 *mask, u8 *val, size_t size) |
| { |
| unsigned int i; |
| |
| for (i = 0; i < size; i++, mask++, val++) |
| *((u8 *)val) = *((u8 *)mask) & *((u8 *)val); |
| } |
| |
| #define MLX5E_FTE_SET(header_p, fld, v) \ |
| MLX5_SET(fte_match_set_lyr_2_4, header_p, fld, v) |
| |
| #define MLX5E_FTE_ADDR_OF(header_p, fld) \ |
| MLX5_ADDR_OF(fte_match_set_lyr_2_4, header_p, fld) |
| |
| static void |
| set_ip4(void *headers_c, void *headers_v, __be32 ip4src_m, |
| __be32 ip4src_v, __be32 ip4dst_m, __be32 ip4dst_v) |
| { |
| if (ip4src_m) { |
| memcpy(MLX5E_FTE_ADDR_OF(headers_v, src_ipv4_src_ipv6.ipv4_layout.ipv4), |
| &ip4src_v, sizeof(ip4src_v)); |
| memcpy(MLX5E_FTE_ADDR_OF(headers_c, src_ipv4_src_ipv6.ipv4_layout.ipv4), |
| &ip4src_m, sizeof(ip4src_m)); |
| } |
| if (ip4dst_m) { |
| memcpy(MLX5E_FTE_ADDR_OF(headers_v, dst_ipv4_dst_ipv6.ipv4_layout.ipv4), |
| &ip4dst_v, sizeof(ip4dst_v)); |
| memcpy(MLX5E_FTE_ADDR_OF(headers_c, dst_ipv4_dst_ipv6.ipv4_layout.ipv4), |
| &ip4dst_m, sizeof(ip4dst_m)); |
| } |
| |
| MLX5E_FTE_SET(headers_c, ethertype, 0xffff); |
| MLX5E_FTE_SET(headers_v, ethertype, ETH_P_IP); |
| } |
| |
| static void |
| set_ip6(void *headers_c, void *headers_v, __be32 ip6src_m[4], |
| __be32 ip6src_v[4], __be32 ip6dst_m[4], __be32 ip6dst_v[4]) |
| { |
| u8 ip6_sz = MLX5_FLD_SZ_BYTES(ipv6_layout, ipv6); |
| |
| if (!ipv6_addr_any((struct in6_addr *)ip6src_m)) { |
| memcpy(MLX5E_FTE_ADDR_OF(headers_v, src_ipv4_src_ipv6.ipv6_layout.ipv6), |
| ip6src_v, ip6_sz); |
| memcpy(MLX5E_FTE_ADDR_OF(headers_c, src_ipv4_src_ipv6.ipv6_layout.ipv6), |
| ip6src_m, ip6_sz); |
| } |
| if (!ipv6_addr_any((struct in6_addr *)ip6dst_m)) { |
| memcpy(MLX5E_FTE_ADDR_OF(headers_v, dst_ipv4_dst_ipv6.ipv6_layout.ipv6), |
| ip6dst_v, ip6_sz); |
| memcpy(MLX5E_FTE_ADDR_OF(headers_c, dst_ipv4_dst_ipv6.ipv6_layout.ipv6), |
| ip6dst_m, ip6_sz); |
| } |
| |
| MLX5E_FTE_SET(headers_c, ethertype, 0xffff); |
| MLX5E_FTE_SET(headers_v, ethertype, ETH_P_IPV6); |
| } |
| |
| static void |
| set_tcp(void *headers_c, void *headers_v, __be16 psrc_m, __be16 psrc_v, |
| __be16 pdst_m, __be16 pdst_v) |
| { |
| if (psrc_m) { |
| MLX5E_FTE_SET(headers_c, tcp_sport, ntohs(psrc_m)); |
| MLX5E_FTE_SET(headers_v, tcp_sport, ntohs(psrc_v)); |
| } |
| if (pdst_m) { |
| MLX5E_FTE_SET(headers_c, tcp_dport, ntohs(pdst_m)); |
| MLX5E_FTE_SET(headers_v, tcp_dport, ntohs(pdst_v)); |
| } |
| |
| MLX5E_FTE_SET(headers_c, ip_protocol, 0xffff); |
| MLX5E_FTE_SET(headers_v, ip_protocol, IPPROTO_TCP); |
| } |
| |
| static void |
| set_udp(void *headers_c, void *headers_v, __be16 psrc_m, __be16 psrc_v, |
| __be16 pdst_m, __be16 pdst_v) |
| { |
| if (psrc_m) { |
| MLX5E_FTE_SET(headers_c, udp_sport, ntohs(psrc_m)); |
| MLX5E_FTE_SET(headers_v, udp_sport, ntohs(psrc_v)); |
| } |
| |
| if (pdst_m) { |
| MLX5E_FTE_SET(headers_c, udp_dport, ntohs(pdst_m)); |
| MLX5E_FTE_SET(headers_v, udp_dport, ntohs(pdst_v)); |
| } |
| |
| MLX5E_FTE_SET(headers_c, ip_protocol, 0xffff); |
| MLX5E_FTE_SET(headers_v, ip_protocol, IPPROTO_UDP); |
| } |
| |
| static void |
| parse_tcp4(void *headers_c, void *headers_v, struct ethtool_rx_flow_spec *fs) |
| { |
| struct ethtool_tcpip4_spec *l4_mask = &fs->m_u.tcp_ip4_spec; |
| struct ethtool_tcpip4_spec *l4_val = &fs->h_u.tcp_ip4_spec; |
| |
| set_ip4(headers_c, headers_v, l4_mask->ip4src, l4_val->ip4src, |
| l4_mask->ip4dst, l4_val->ip4dst); |
| |
| set_tcp(headers_c, headers_v, l4_mask->psrc, l4_val->psrc, |
| l4_mask->pdst, l4_val->pdst); |
| } |
| |
| static void |
| parse_udp4(void *headers_c, void *headers_v, struct ethtool_rx_flow_spec *fs) |
| { |
| struct ethtool_tcpip4_spec *l4_mask = &fs->m_u.udp_ip4_spec; |
| struct ethtool_tcpip4_spec *l4_val = &fs->h_u.udp_ip4_spec; |
| |
| set_ip4(headers_c, headers_v, l4_mask->ip4src, l4_val->ip4src, |
| l4_mask->ip4dst, l4_val->ip4dst); |
| |
| set_udp(headers_c, headers_v, l4_mask->psrc, l4_val->psrc, |
| l4_mask->pdst, l4_val->pdst); |
| } |
| |
| static void |
| parse_ip4(void *headers_c, void *headers_v, struct ethtool_rx_flow_spec *fs) |
| { |
| struct ethtool_usrip4_spec *l3_mask = &fs->m_u.usr_ip4_spec; |
| struct ethtool_usrip4_spec *l3_val = &fs->h_u.usr_ip4_spec; |
| |
| set_ip4(headers_c, headers_v, l3_mask->ip4src, l3_val->ip4src, |
| l3_mask->ip4dst, l3_val->ip4dst); |
| |
| if (l3_mask->proto) { |
| MLX5E_FTE_SET(headers_c, ip_protocol, l3_mask->proto); |
| MLX5E_FTE_SET(headers_v, ip_protocol, l3_val->proto); |
| } |
| } |
| |
| static void |
| parse_ip6(void *headers_c, void *headers_v, struct ethtool_rx_flow_spec *fs) |
| { |
| struct ethtool_usrip6_spec *l3_mask = &fs->m_u.usr_ip6_spec; |
| struct ethtool_usrip6_spec *l3_val = &fs->h_u.usr_ip6_spec; |
| |
| set_ip6(headers_c, headers_v, l3_mask->ip6src, |
| l3_val->ip6src, l3_mask->ip6dst, l3_val->ip6dst); |
| |
| if (l3_mask->l4_proto) { |
| MLX5E_FTE_SET(headers_c, ip_protocol, l3_mask->l4_proto); |
| MLX5E_FTE_SET(headers_v, ip_protocol, l3_val->l4_proto); |
| } |
| } |
| |
| static void |
| parse_tcp6(void *headers_c, void *headers_v, struct ethtool_rx_flow_spec *fs) |
| { |
| struct ethtool_tcpip6_spec *l4_mask = &fs->m_u.tcp_ip6_spec; |
| struct ethtool_tcpip6_spec *l4_val = &fs->h_u.tcp_ip6_spec; |
| |
| set_ip6(headers_c, headers_v, l4_mask->ip6src, |
| l4_val->ip6src, l4_mask->ip6dst, l4_val->ip6dst); |
| |
| set_tcp(headers_c, headers_v, l4_mask->psrc, l4_val->psrc, |
| l4_mask->pdst, l4_val->pdst); |
| } |
| |
| static void |
| parse_udp6(void *headers_c, void *headers_v, struct ethtool_rx_flow_spec *fs) |
| { |
| struct ethtool_tcpip6_spec *l4_mask = &fs->m_u.udp_ip6_spec; |
| struct ethtool_tcpip6_spec *l4_val = &fs->h_u.udp_ip6_spec; |
| |
| set_ip6(headers_c, headers_v, l4_mask->ip6src, |
| l4_val->ip6src, l4_mask->ip6dst, l4_val->ip6dst); |
| |
| set_udp(headers_c, headers_v, l4_mask->psrc, l4_val->psrc, |
| l4_mask->pdst, l4_val->pdst); |
| } |
| |
| static void |
| parse_ether(void *headers_c, void *headers_v, struct ethtool_rx_flow_spec *fs) |
| { |
| struct ethhdr *eth_mask = &fs->m_u.ether_spec; |
| struct ethhdr *eth_val = &fs->h_u.ether_spec; |
| |
| mask_spec((u8 *)eth_mask, (u8 *)eth_val, sizeof(*eth_mask)); |
| ether_addr_copy(MLX5E_FTE_ADDR_OF(headers_c, smac_47_16), eth_mask->h_source); |
| ether_addr_copy(MLX5E_FTE_ADDR_OF(headers_v, smac_47_16), eth_val->h_source); |
| ether_addr_copy(MLX5E_FTE_ADDR_OF(headers_c, dmac_47_16), eth_mask->h_dest); |
| ether_addr_copy(MLX5E_FTE_ADDR_OF(headers_v, dmac_47_16), eth_val->h_dest); |
| MLX5E_FTE_SET(headers_c, ethertype, ntohs(eth_mask->h_proto)); |
| MLX5E_FTE_SET(headers_v, ethertype, ntohs(eth_val->h_proto)); |
| } |
| |
| static void |
| set_cvlan(void *headers_c, void *headers_v, __be16 vlan_tci) |
| { |
| MLX5E_FTE_SET(headers_c, cvlan_tag, 1); |
| MLX5E_FTE_SET(headers_v, cvlan_tag, 1); |
| MLX5E_FTE_SET(headers_c, first_vid, 0xfff); |
| MLX5E_FTE_SET(headers_v, first_vid, ntohs(vlan_tci)); |
| } |
| |
| static void |
| set_dmac(void *headers_c, void *headers_v, |
| unsigned char m_dest[ETH_ALEN], unsigned char v_dest[ETH_ALEN]) |
| { |
| ether_addr_copy(MLX5E_FTE_ADDR_OF(headers_c, dmac_47_16), m_dest); |
| ether_addr_copy(MLX5E_FTE_ADDR_OF(headers_v, dmac_47_16), v_dest); |
| } |
| |
| static int set_flow_attrs(u32 *match_c, u32 *match_v, |
| struct ethtool_rx_flow_spec *fs) |
| { |
| void *outer_headers_c = MLX5_ADDR_OF(fte_match_param, match_c, |
| outer_headers); |
| void *outer_headers_v = MLX5_ADDR_OF(fte_match_param, match_v, |
| outer_headers); |
| u32 flow_type = fs->flow_type & ~(FLOW_EXT | FLOW_MAC_EXT); |
| |
| switch (flow_type) { |
| case TCP_V4_FLOW: |
| parse_tcp4(outer_headers_c, outer_headers_v, fs); |
| break; |
| case UDP_V4_FLOW: |
| parse_udp4(outer_headers_c, outer_headers_v, fs); |
| break; |
| case IP_USER_FLOW: |
| parse_ip4(outer_headers_c, outer_headers_v, fs); |
| break; |
| case TCP_V6_FLOW: |
| parse_tcp6(outer_headers_c, outer_headers_v, fs); |
| break; |
| case UDP_V6_FLOW: |
| parse_udp6(outer_headers_c, outer_headers_v, fs); |
| break; |
| case IPV6_USER_FLOW: |
| parse_ip6(outer_headers_c, outer_headers_v, fs); |
| break; |
| case ETHER_FLOW: |
| parse_ether(outer_headers_c, outer_headers_v, fs); |
| break; |
| default: |
| return -EINVAL; |
| } |
| |
| if ((fs->flow_type & FLOW_EXT) && |
| (fs->m_ext.vlan_tci & cpu_to_be16(VLAN_VID_MASK))) |
| set_cvlan(outer_headers_c, outer_headers_v, fs->h_ext.vlan_tci); |
| |
| if (fs->flow_type & FLOW_MAC_EXT && |
| !is_zero_ether_addr(fs->m_ext.h_dest)) { |
| mask_spec(fs->m_ext.h_dest, fs->h_ext.h_dest, ETH_ALEN); |
| set_dmac(outer_headers_c, outer_headers_v, fs->m_ext.h_dest, |
| fs->h_ext.h_dest); |
| } |
| |
| return 0; |
| } |
| |
| static void add_rule_to_list(struct mlx5e_priv *priv, |
| struct mlx5e_ethtool_rule *rule) |
| { |
| struct mlx5e_ethtool_rule *iter; |
| struct list_head *head = &priv->fs.ethtool.rules; |
| |
| list_for_each_entry(iter, &priv->fs.ethtool.rules, list) { |
| if (iter->flow_spec.location > rule->flow_spec.location) |
| break; |
| head = &iter->list; |
| } |
| priv->fs.ethtool.tot_num_rules++; |
| list_add(&rule->list, head); |
| } |
| |
| static bool outer_header_zero(u32 *match_criteria) |
| { |
| int size = MLX5_FLD_SZ_BYTES(fte_match_param, outer_headers); |
| char *outer_headers_c = MLX5_ADDR_OF(fte_match_param, match_criteria, |
| outer_headers); |
| |
| return outer_headers_c[0] == 0 && !memcmp(outer_headers_c, |
| outer_headers_c + 1, |
| size - 1); |
| } |
| |
| static struct mlx5_flow_handle * |
| add_ethtool_flow_rule(struct mlx5e_priv *priv, |
| struct mlx5_flow_table *ft, |
| struct ethtool_rx_flow_spec *fs) |
| { |
| struct mlx5_flow_act flow_act = { .flags = FLOW_ACT_NO_APPEND }; |
| struct mlx5_flow_destination *dst = NULL; |
| struct mlx5_flow_handle *rule; |
| struct mlx5_flow_spec *spec; |
| int err = 0; |
| |
| spec = kvzalloc(sizeof(*spec), GFP_KERNEL); |
| if (!spec) |
| return ERR_PTR(-ENOMEM); |
| err = set_flow_attrs(spec->match_criteria, spec->match_value, |
| fs); |
| if (err) |
| goto free; |
| |
| if (fs->ring_cookie == RX_CLS_FLOW_DISC) { |
| flow_act.action = MLX5_FLOW_CONTEXT_ACTION_DROP; |
| } else { |
| struct mlx5e_params *params = &priv->channels.params; |
| enum mlx5e_rq_group group; |
| struct mlx5e_tir *tir; |
| u16 ix; |
| |
| mlx5e_qid_get_ch_and_group(params, fs->ring_cookie, &ix, &group); |
| tir = group == MLX5E_RQ_GROUP_XSK ? priv->xsk_tir : priv->direct_tir; |
| |
| dst = kzalloc(sizeof(*dst), GFP_KERNEL); |
| if (!dst) { |
| err = -ENOMEM; |
| goto free; |
| } |
| |
| dst->type = MLX5_FLOW_DESTINATION_TYPE_TIR; |
| dst->tir_num = tir[ix].tirn; |
| flow_act.action = MLX5_FLOW_CONTEXT_ACTION_FWD_DEST; |
| } |
| |
| spec->match_criteria_enable = (!outer_header_zero(spec->match_criteria)); |
| spec->flow_context.flow_tag = MLX5_FS_DEFAULT_FLOW_TAG; |
| rule = mlx5_add_flow_rules(ft, spec, &flow_act, dst, dst ? 1 : 0); |
| if (IS_ERR(rule)) { |
| err = PTR_ERR(rule); |
| netdev_err(priv->netdev, "%s: failed to add ethtool steering rule: %d\n", |
| __func__, err); |
| goto free; |
| } |
| free: |
| kvfree(spec); |
| kfree(dst); |
| return err ? ERR_PTR(err) : rule; |
| } |
| |
| static void del_ethtool_rule(struct mlx5e_priv *priv, |
| struct mlx5e_ethtool_rule *eth_rule) |
| { |
| if (eth_rule->rule) |
| mlx5_del_flow_rules(eth_rule->rule); |
| list_del(ð_rule->list); |
| priv->fs.ethtool.tot_num_rules--; |
| put_flow_table(eth_rule->eth_ft); |
| kfree(eth_rule); |
| } |
| |
| static struct mlx5e_ethtool_rule *find_ethtool_rule(struct mlx5e_priv *priv, |
| int location) |
| { |
| struct mlx5e_ethtool_rule *iter; |
| |
| list_for_each_entry(iter, &priv->fs.ethtool.rules, list) { |
| if (iter->flow_spec.location == location) |
| return iter; |
| } |
| return NULL; |
| } |
| |
| static struct mlx5e_ethtool_rule *get_ethtool_rule(struct mlx5e_priv *priv, |
| int location) |
| { |
| struct mlx5e_ethtool_rule *eth_rule; |
| |
| eth_rule = find_ethtool_rule(priv, location); |
| if (eth_rule) |
| del_ethtool_rule(priv, eth_rule); |
| |
| eth_rule = kzalloc(sizeof(*eth_rule), GFP_KERNEL); |
| if (!eth_rule) |
| return ERR_PTR(-ENOMEM); |
| |
| add_rule_to_list(priv, eth_rule); |
| return eth_rule; |
| } |
| |
| #define MAX_NUM_OF_ETHTOOL_RULES BIT(10) |
| |
| #define all_ones(field) (field == (__force typeof(field))-1) |
| #define all_zeros_or_all_ones(field) \ |
| ((field) == 0 || (field) == (__force typeof(field))-1) |
| |
| static int validate_ethter(struct ethtool_rx_flow_spec *fs) |
| { |
| struct ethhdr *eth_mask = &fs->m_u.ether_spec; |
| int ntuples = 0; |
| |
| if (!is_zero_ether_addr(eth_mask->h_dest)) |
| ntuples++; |
| if (!is_zero_ether_addr(eth_mask->h_source)) |
| ntuples++; |
| if (eth_mask->h_proto) |
| ntuples++; |
| return ntuples; |
| } |
| |
| static int validate_tcpudp4(struct ethtool_rx_flow_spec *fs) |
| { |
| struct ethtool_tcpip4_spec *l4_mask = &fs->m_u.tcp_ip4_spec; |
| int ntuples = 0; |
| |
| if (l4_mask->tos) |
| return -EINVAL; |
| |
| if (l4_mask->ip4src) |
| ntuples++; |
| if (l4_mask->ip4dst) |
| ntuples++; |
| if (l4_mask->psrc) |
| ntuples++; |
| if (l4_mask->pdst) |
| ntuples++; |
| /* Flow is TCP/UDP */ |
| return ++ntuples; |
| } |
| |
| static int validate_ip4(struct ethtool_rx_flow_spec *fs) |
| { |
| struct ethtool_usrip4_spec *l3_mask = &fs->m_u.usr_ip4_spec; |
| int ntuples = 0; |
| |
| if (l3_mask->l4_4_bytes || l3_mask->tos || |
| fs->h_u.usr_ip4_spec.ip_ver != ETH_RX_NFC_IP4) |
| return -EINVAL; |
| if (l3_mask->ip4src) |
| ntuples++; |
| if (l3_mask->ip4dst) |
| ntuples++; |
| if (l3_mask->proto) |
| ntuples++; |
| /* Flow is IPv4 */ |
| return ++ntuples; |
| } |
| |
| static int validate_ip6(struct ethtool_rx_flow_spec *fs) |
| { |
| struct ethtool_usrip6_spec *l3_mask = &fs->m_u.usr_ip6_spec; |
| int ntuples = 0; |
| |
| if (l3_mask->l4_4_bytes || l3_mask->tclass) |
| return -EINVAL; |
| if (!ipv6_addr_any((struct in6_addr *)l3_mask->ip6src)) |
| ntuples++; |
| |
| if (!ipv6_addr_any((struct in6_addr *)l3_mask->ip6dst)) |
| ntuples++; |
| if (l3_mask->l4_proto) |
| ntuples++; |
| /* Flow is IPv6 */ |
| return ++ntuples; |
| } |
| |
| static int validate_tcpudp6(struct ethtool_rx_flow_spec *fs) |
| { |
| struct ethtool_tcpip6_spec *l4_mask = &fs->m_u.tcp_ip6_spec; |
| int ntuples = 0; |
| |
| if (l4_mask->tclass) |
| return -EINVAL; |
| |
| if (!ipv6_addr_any((struct in6_addr *)l4_mask->ip6src)) |
| ntuples++; |
| |
| if (!ipv6_addr_any((struct in6_addr *)l4_mask->ip6dst)) |
| ntuples++; |
| |
| if (l4_mask->psrc) |
| ntuples++; |
| if (l4_mask->pdst) |
| ntuples++; |
| /* Flow is TCP/UDP */ |
| return ++ntuples; |
| } |
| |
| static int validate_vlan(struct ethtool_rx_flow_spec *fs) |
| { |
| if (fs->m_ext.vlan_etype || |
| fs->m_ext.vlan_tci != cpu_to_be16(VLAN_VID_MASK)) |
| return -EINVAL; |
| |
| if (fs->m_ext.vlan_tci && |
| (be16_to_cpu(fs->h_ext.vlan_tci) >= VLAN_N_VID)) |
| return -EINVAL; |
| |
| return 1; |
| } |
| |
| static int validate_flow(struct mlx5e_priv *priv, |
| struct ethtool_rx_flow_spec *fs) |
| { |
| int num_tuples = 0; |
| int ret = 0; |
| |
| if (fs->location >= MAX_NUM_OF_ETHTOOL_RULES) |
| return -ENOSPC; |
| |
| if (fs->ring_cookie != RX_CLS_FLOW_DISC) |
| if (!mlx5e_qid_validate(priv->profile, &priv->channels.params, |
| fs->ring_cookie)) |
| return -EINVAL; |
| |
| switch (fs->flow_type & ~(FLOW_EXT | FLOW_MAC_EXT)) { |
| case ETHER_FLOW: |
| num_tuples += validate_ethter(fs); |
| break; |
| case TCP_V4_FLOW: |
| case UDP_V4_FLOW: |
| ret = validate_tcpudp4(fs); |
| if (ret < 0) |
| return ret; |
| num_tuples += ret; |
| break; |
| case IP_USER_FLOW: |
| ret = validate_ip4(fs); |
| if (ret < 0) |
| return ret; |
| num_tuples += ret; |
| break; |
| case TCP_V6_FLOW: |
| case UDP_V6_FLOW: |
| ret = validate_tcpudp6(fs); |
| if (ret < 0) |
| return ret; |
| num_tuples += ret; |
| break; |
| case IPV6_USER_FLOW: |
| ret = validate_ip6(fs); |
| if (ret < 0) |
| return ret; |
| num_tuples += ret; |
| break; |
| default: |
| return -ENOTSUPP; |
| } |
| if ((fs->flow_type & FLOW_EXT)) { |
| ret = validate_vlan(fs); |
| if (ret < 0) |
| return ret; |
| num_tuples += ret; |
| } |
| |
| if (fs->flow_type & FLOW_MAC_EXT && |
| !is_zero_ether_addr(fs->m_ext.h_dest)) |
| num_tuples++; |
| |
| return num_tuples; |
| } |
| |
| static int |
| mlx5e_ethtool_flow_replace(struct mlx5e_priv *priv, |
| struct ethtool_rx_flow_spec *fs) |
| { |
| struct mlx5e_ethtool_table *eth_ft; |
| struct mlx5e_ethtool_rule *eth_rule; |
| struct mlx5_flow_handle *rule; |
| int num_tuples; |
| int err; |
| |
| num_tuples = validate_flow(priv, fs); |
| if (num_tuples <= 0) { |
| netdev_warn(priv->netdev, "%s: flow is not valid %d\n", |
| __func__, num_tuples); |
| return num_tuples; |
| } |
| |
| eth_ft = get_flow_table(priv, fs, num_tuples); |
| if (IS_ERR(eth_ft)) |
| return PTR_ERR(eth_ft); |
| |
| eth_rule = get_ethtool_rule(priv, fs->location); |
| if (IS_ERR(eth_rule)) { |
| put_flow_table(eth_ft); |
| return PTR_ERR(eth_rule); |
| } |
| |
| eth_rule->flow_spec = *fs; |
| eth_rule->eth_ft = eth_ft; |
| if (!eth_ft->ft) { |
| err = -EINVAL; |
| goto del_ethtool_rule; |
| } |
| rule = add_ethtool_flow_rule(priv, eth_ft->ft, fs); |
| if (IS_ERR(rule)) { |
| err = PTR_ERR(rule); |
| goto del_ethtool_rule; |
| } |
| |
| eth_rule->rule = rule; |
| |
| return 0; |
| |
| del_ethtool_rule: |
| del_ethtool_rule(priv, eth_rule); |
| |
| return err; |
| } |
| |
| static int |
| mlx5e_ethtool_flow_remove(struct mlx5e_priv *priv, int location) |
| { |
| struct mlx5e_ethtool_rule *eth_rule; |
| int err = 0; |
| |
| if (location >= MAX_NUM_OF_ETHTOOL_RULES) |
| return -ENOSPC; |
| |
| eth_rule = find_ethtool_rule(priv, location); |
| if (!eth_rule) { |
| err = -ENOENT; |
| goto out; |
| } |
| |
| del_ethtool_rule(priv, eth_rule); |
| out: |
| return err; |
| } |
| |
| static int |
| mlx5e_ethtool_get_flow(struct mlx5e_priv *priv, |
| struct ethtool_rxnfc *info, int location) |
| { |
| struct mlx5e_ethtool_rule *eth_rule; |
| |
| if (location < 0 || location >= MAX_NUM_OF_ETHTOOL_RULES) |
| return -EINVAL; |
| |
| list_for_each_entry(eth_rule, &priv->fs.ethtool.rules, list) { |
| if (eth_rule->flow_spec.location == location) { |
| info->fs = eth_rule->flow_spec; |
| return 0; |
| } |
| } |
| |
| return -ENOENT; |
| } |
| |
| static int |
| mlx5e_ethtool_get_all_flows(struct mlx5e_priv *priv, |
| struct ethtool_rxnfc *info, u32 *rule_locs) |
| { |
| int location = 0; |
| int idx = 0; |
| int err = 0; |
| |
| info->data = MAX_NUM_OF_ETHTOOL_RULES; |
| while ((!err || err == -ENOENT) && idx < info->rule_cnt) { |
| err = mlx5e_ethtool_get_flow(priv, info, location); |
| if (!err) |
| rule_locs[idx++] = location; |
| location++; |
| } |
| return err; |
| } |
| |
| void mlx5e_ethtool_cleanup_steering(struct mlx5e_priv *priv) |
| { |
| struct mlx5e_ethtool_rule *iter; |
| struct mlx5e_ethtool_rule *temp; |
| |
| list_for_each_entry_safe(iter, temp, &priv->fs.ethtool.rules, list) |
| del_ethtool_rule(priv, iter); |
| } |
| |
| void mlx5e_ethtool_init_steering(struct mlx5e_priv *priv) |
| { |
| INIT_LIST_HEAD(&priv->fs.ethtool.rules); |
| } |
| |
| static enum mlx5e_traffic_types flow_type_to_traffic_type(u32 flow_type) |
| { |
| switch (flow_type) { |
| case TCP_V4_FLOW: |
| return MLX5E_TT_IPV4_TCP; |
| case TCP_V6_FLOW: |
| return MLX5E_TT_IPV6_TCP; |
| case UDP_V4_FLOW: |
| return MLX5E_TT_IPV4_UDP; |
| case UDP_V6_FLOW: |
| return MLX5E_TT_IPV6_UDP; |
| case AH_V4_FLOW: |
| return MLX5E_TT_IPV4_IPSEC_AH; |
| case AH_V6_FLOW: |
| return MLX5E_TT_IPV6_IPSEC_AH; |
| case ESP_V4_FLOW: |
| return MLX5E_TT_IPV4_IPSEC_ESP; |
| case ESP_V6_FLOW: |
| return MLX5E_TT_IPV6_IPSEC_ESP; |
| case IPV4_FLOW: |
| return MLX5E_TT_IPV4; |
| case IPV6_FLOW: |
| return MLX5E_TT_IPV6; |
| default: |
| return MLX5E_NUM_INDIR_TIRS; |
| } |
| } |
| |
| static int mlx5e_set_rss_hash_opt(struct mlx5e_priv *priv, |
| struct ethtool_rxnfc *nfc) |
| { |
| int inlen = MLX5_ST_SZ_BYTES(modify_tir_in); |
| enum mlx5e_traffic_types tt; |
| u8 rx_hash_field = 0; |
| void *in; |
| |
| tt = flow_type_to_traffic_type(nfc->flow_type); |
| if (tt == MLX5E_NUM_INDIR_TIRS) |
| return -EINVAL; |
| |
| /* RSS does not support anything other than hashing to queues |
| * on src IP, dest IP, TCP/UDP src port and TCP/UDP dest |
| * port. |
| */ |
| if (nfc->flow_type != TCP_V4_FLOW && |
| nfc->flow_type != TCP_V6_FLOW && |
| nfc->flow_type != UDP_V4_FLOW && |
| nfc->flow_type != UDP_V6_FLOW) |
| return -EOPNOTSUPP; |
| |
| if (nfc->data & ~(RXH_IP_SRC | RXH_IP_DST | |
| RXH_L4_B_0_1 | RXH_L4_B_2_3)) |
| return -EOPNOTSUPP; |
| |
| if (nfc->data & RXH_IP_SRC) |
| rx_hash_field |= MLX5_HASH_FIELD_SEL_SRC_IP; |
| if (nfc->data & RXH_IP_DST) |
| rx_hash_field |= MLX5_HASH_FIELD_SEL_DST_IP; |
| if (nfc->data & RXH_L4_B_0_1) |
| rx_hash_field |= MLX5_HASH_FIELD_SEL_L4_SPORT; |
| if (nfc->data & RXH_L4_B_2_3) |
| rx_hash_field |= MLX5_HASH_FIELD_SEL_L4_DPORT; |
| |
| in = kvzalloc(inlen, GFP_KERNEL); |
| if (!in) |
| return -ENOMEM; |
| |
| mutex_lock(&priv->state_lock); |
| |
| if (rx_hash_field == priv->rss_params.rx_hash_fields[tt]) |
| goto out; |
| |
| priv->rss_params.rx_hash_fields[tt] = rx_hash_field; |
| mlx5e_modify_tirs_hash(priv, in); |
| |
| out: |
| mutex_unlock(&priv->state_lock); |
| kvfree(in); |
| return 0; |
| } |
| |
| static int mlx5e_get_rss_hash_opt(struct mlx5e_priv *priv, |
| struct ethtool_rxnfc *nfc) |
| { |
| enum mlx5e_traffic_types tt; |
| u32 hash_field = 0; |
| |
| tt = flow_type_to_traffic_type(nfc->flow_type); |
| if (tt == MLX5E_NUM_INDIR_TIRS) |
| return -EINVAL; |
| |
| hash_field = priv->rss_params.rx_hash_fields[tt]; |
| nfc->data = 0; |
| |
| if (hash_field & MLX5_HASH_FIELD_SEL_SRC_IP) |
| nfc->data |= RXH_IP_SRC; |
| if (hash_field & MLX5_HASH_FIELD_SEL_DST_IP) |
| nfc->data |= RXH_IP_DST; |
| if (hash_field & MLX5_HASH_FIELD_SEL_L4_SPORT) |
| nfc->data |= RXH_L4_B_0_1; |
| if (hash_field & MLX5_HASH_FIELD_SEL_L4_DPORT) |
| nfc->data |= RXH_L4_B_2_3; |
| |
| return 0; |
| } |
| |
| int mlx5e_ethtool_set_rxnfc(struct net_device *dev, struct ethtool_rxnfc *cmd) |
| { |
| struct mlx5e_priv *priv = netdev_priv(dev); |
| int err = 0; |
| |
| switch (cmd->cmd) { |
| case ETHTOOL_SRXCLSRLINS: |
| err = mlx5e_ethtool_flow_replace(priv, &cmd->fs); |
| break; |
| case ETHTOOL_SRXCLSRLDEL: |
| err = mlx5e_ethtool_flow_remove(priv, cmd->fs.location); |
| break; |
| case ETHTOOL_SRXFH: |
| err = mlx5e_set_rss_hash_opt(priv, cmd); |
| break; |
| default: |
| err = -EOPNOTSUPP; |
| break; |
| } |
| |
| return err; |
| } |
| |
| int mlx5e_ethtool_get_rxnfc(struct net_device *dev, |
| struct ethtool_rxnfc *info, u32 *rule_locs) |
| { |
| struct mlx5e_priv *priv = netdev_priv(dev); |
| int err = 0; |
| |
| switch (info->cmd) { |
| case ETHTOOL_GRXCLSRLCNT: |
| info->rule_cnt = priv->fs.ethtool.tot_num_rules; |
| break; |
| case ETHTOOL_GRXCLSRULE: |
| err = mlx5e_ethtool_get_flow(priv, info, info->fs.location); |
| break; |
| case ETHTOOL_GRXCLSRLALL: |
| err = mlx5e_ethtool_get_all_flows(priv, info, rule_locs); |
| break; |
| case ETHTOOL_GRXFH: |
| err = mlx5e_get_rss_hash_opt(priv, info); |
| break; |
| default: |
| err = -EOPNOTSUPP; |
| break; |
| } |
| |
| return err; |
| } |
| |