cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

ib_virt.c (6373B)


      1/*
      2 * Copyright (c) 2016, Mellanox Technologies. All rights reserved.
      3 *
      4 * This software is available to you under a choice of one of two
      5 * licenses.  You may choose to be licensed under the terms of the GNU
      6 * General Public License (GPL) Version 2, available from the file
      7 * COPYING in the main directory of this source tree, or the
      8 * OpenIB.org BSD license below:
      9 *
     10 *     Redistribution and use in source and binary forms, with or
     11 *     without modification, are permitted provided that the following
     12 *     conditions are met:
     13 *
     14 *      - Redistributions of source code must retain the above
     15 *        copyright notice, this list of conditions and the following
     16 *        disclaimer.
     17 *
     18 *      - Redistributions in binary form must reproduce the above
     19 *        copyright notice, this list of conditions and the following
     20 *        disclaimer in the documentation and/or other materials
     21 *        provided with the distribution.
     22 *
     23 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
     24 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
     25 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
     26 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
     27 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
     28 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
     29 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
     30 * SOFTWARE.
     31 */
     32
     33#include <linux/mlx5/vport.h>
     34#include "mlx5_ib.h"
     35
     36static inline u32 mlx_to_net_policy(enum port_state_policy mlx_policy)
     37{
     38	switch (mlx_policy) {
     39	case MLX5_POLICY_DOWN:
     40		return IFLA_VF_LINK_STATE_DISABLE;
     41	case MLX5_POLICY_UP:
     42		return IFLA_VF_LINK_STATE_ENABLE;
     43	case MLX5_POLICY_FOLLOW:
     44		return IFLA_VF_LINK_STATE_AUTO;
     45	default:
     46		return __IFLA_VF_LINK_STATE_MAX;
     47	}
     48}
     49
     50int mlx5_ib_get_vf_config(struct ib_device *device, int vf, u32 port,
     51			  struct ifla_vf_info *info)
     52{
     53	struct mlx5_ib_dev *dev = to_mdev(device);
     54	struct mlx5_core_dev *mdev = dev->mdev;
     55	struct mlx5_hca_vport_context *rep;
     56	int err;
     57
     58	rep = kzalloc(sizeof(*rep), GFP_KERNEL);
     59	if (!rep)
     60		return -ENOMEM;
     61
     62	err = mlx5_query_hca_vport_context(mdev, 1, 1,  vf + 1, rep);
     63	if (err) {
     64		mlx5_ib_warn(dev, "failed to query port policy for vf %d (%d)\n",
     65			     vf, err);
     66		goto free;
     67	}
     68	memset(info, 0, sizeof(*info));
     69	info->linkstate = mlx_to_net_policy(rep->policy);
     70	if (info->linkstate == __IFLA_VF_LINK_STATE_MAX)
     71		err = -EINVAL;
     72
     73free:
     74	kfree(rep);
     75	return err;
     76}
     77
     78static inline enum port_state_policy net_to_mlx_policy(int policy)
     79{
     80	switch (policy) {
     81	case IFLA_VF_LINK_STATE_DISABLE:
     82		return MLX5_POLICY_DOWN;
     83	case IFLA_VF_LINK_STATE_ENABLE:
     84		return MLX5_POLICY_UP;
     85	case IFLA_VF_LINK_STATE_AUTO:
     86		return MLX5_POLICY_FOLLOW;
     87	default:
     88		return MLX5_POLICY_INVALID;
     89	}
     90}
     91
     92int mlx5_ib_set_vf_link_state(struct ib_device *device, int vf,
     93			      u32 port, int state)
     94{
     95	struct mlx5_ib_dev *dev = to_mdev(device);
     96	struct mlx5_core_dev *mdev = dev->mdev;
     97	struct mlx5_hca_vport_context *in;
     98	struct mlx5_vf_context *vfs_ctx = mdev->priv.sriov.vfs_ctx;
     99	int err;
    100
    101	in = kzalloc(sizeof(*in), GFP_KERNEL);
    102	if (!in)
    103		return -ENOMEM;
    104
    105	in->policy = net_to_mlx_policy(state);
    106	if (in->policy == MLX5_POLICY_INVALID) {
    107		err = -EINVAL;
    108		goto out;
    109	}
    110	in->field_select = MLX5_HCA_VPORT_SEL_STATE_POLICY;
    111	err = mlx5_core_modify_hca_vport_context(mdev, 1, 1, vf + 1, in);
    112	if (!err)
    113		vfs_ctx[vf].policy = in->policy;
    114
    115out:
    116	kfree(in);
    117	return err;
    118}
    119
    120int mlx5_ib_get_vf_stats(struct ib_device *device, int vf,
    121			 u32 port, struct ifla_vf_stats *stats)
    122{
    123	int out_sz = MLX5_ST_SZ_BYTES(query_vport_counter_out);
    124	struct mlx5_core_dev *mdev;
    125	struct mlx5_ib_dev *dev;
    126	void *out;
    127	int err;
    128
    129	dev = to_mdev(device);
    130	mdev = dev->mdev;
    131
    132	out = kzalloc(out_sz, GFP_KERNEL);
    133	if (!out)
    134		return -ENOMEM;
    135
    136	err = mlx5_core_query_vport_counter(mdev, true, vf, port, out);
    137	if (err)
    138		goto ex;
    139
    140	stats->rx_packets = MLX5_GET64_PR(query_vport_counter_out, out, received_ib_unicast.packets);
    141	stats->tx_packets = MLX5_GET64_PR(query_vport_counter_out, out, transmitted_ib_unicast.packets);
    142	stats->rx_bytes = MLX5_GET64_PR(query_vport_counter_out, out, received_ib_unicast.octets);
    143	stats->tx_bytes = MLX5_GET64_PR(query_vport_counter_out, out, transmitted_ib_unicast.octets);
    144	stats->multicast = MLX5_GET64_PR(query_vport_counter_out, out, received_ib_multicast.packets);
    145
    146ex:
    147	kfree(out);
    148	return err;
    149}
    150
    151static int set_vf_node_guid(struct ib_device *device, int vf, u32 port,
    152			    u64 guid)
    153{
    154	struct mlx5_ib_dev *dev = to_mdev(device);
    155	struct mlx5_core_dev *mdev = dev->mdev;
    156	struct mlx5_hca_vport_context *in;
    157	struct mlx5_vf_context *vfs_ctx = mdev->priv.sriov.vfs_ctx;
    158	int err;
    159
    160	in = kzalloc(sizeof(*in), GFP_KERNEL);
    161	if (!in)
    162		return -ENOMEM;
    163
    164	in->field_select = MLX5_HCA_VPORT_SEL_NODE_GUID;
    165	in->node_guid = guid;
    166	err = mlx5_core_modify_hca_vport_context(mdev, 1, 1, vf + 1, in);
    167	if (!err) {
    168		vfs_ctx[vf].node_guid = guid;
    169		vfs_ctx[vf].node_guid_valid = 1;
    170	}
    171	kfree(in);
    172	return err;
    173}
    174
    175static int set_vf_port_guid(struct ib_device *device, int vf, u32 port,
    176			    u64 guid)
    177{
    178	struct mlx5_ib_dev *dev = to_mdev(device);
    179	struct mlx5_core_dev *mdev = dev->mdev;
    180	struct mlx5_hca_vport_context *in;
    181	struct mlx5_vf_context *vfs_ctx = mdev->priv.sriov.vfs_ctx;
    182	int err;
    183
    184	in = kzalloc(sizeof(*in), GFP_KERNEL);
    185	if (!in)
    186		return -ENOMEM;
    187
    188	in->field_select = MLX5_HCA_VPORT_SEL_PORT_GUID;
    189	in->port_guid = guid;
    190	err = mlx5_core_modify_hca_vport_context(mdev, 1, 1, vf + 1, in);
    191	if (!err) {
    192		vfs_ctx[vf].port_guid = guid;
    193		vfs_ctx[vf].port_guid_valid = 1;
    194	}
    195	kfree(in);
    196	return err;
    197}
    198
    199int mlx5_ib_set_vf_guid(struct ib_device *device, int vf, u32 port,
    200			u64 guid, int type)
    201{
    202	if (type == IFLA_VF_IB_NODE_GUID)
    203		return set_vf_node_guid(device, vf, port, guid);
    204	else if (type == IFLA_VF_IB_PORT_GUID)
    205		return set_vf_port_guid(device, vf, port, guid);
    206
    207	return -EINVAL;
    208}
    209
    210int mlx5_ib_get_vf_guid(struct ib_device *device, int vf, u32 port,
    211			struct ifla_vf_guid *node_guid,
    212			struct ifla_vf_guid *port_guid)
    213{
    214	struct mlx5_ib_dev *dev = to_mdev(device);
    215	struct mlx5_core_dev *mdev = dev->mdev;
    216	struct mlx5_vf_context *vfs_ctx = mdev->priv.sriov.vfs_ctx;
    217
    218	node_guid->guid =
    219		vfs_ctx[vf].node_guid_valid ? vfs_ctx[vf].node_guid : 0;
    220	port_guid->guid =
    221		vfs_ctx[vf].port_guid_valid ? vfs_ctx[vf].port_guid : 0;
    222
    223	return 0;
    224}