Contributors: 9
Author |
Tokens |
Token Proportion |
Commits |
Commit Proportion |
Bodong Wang |
192 |
44.65% |
2 |
12.50% |
Parav Pandit |
139 |
32.33% |
2 |
12.50% |
Eli Cohen |
58 |
13.49% |
5 |
31.25% |
Maher Sanalla |
13 |
3.02% |
2 |
12.50% |
Aviv Heller |
12 |
2.79% |
1 |
6.25% |
Leon Romanovsky |
6 |
1.40% |
1 |
6.25% |
Daniel Jurgens |
6 |
1.40% |
1 |
6.25% |
Shay Drory |
3 |
0.70% |
1 |
6.25% |
Qinglang Miao |
1 |
0.23% |
1 |
6.25% |
Total |
430 |
|
16 |
|
/* SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB */
/* Copyright (c) 2019 Mellanox Technologies. */
#include "ecpf.h"
bool mlx5_read_embedded_cpu(struct mlx5_core_dev *dev)
{
return (ioread32be(&dev->iseg->initializing) >> MLX5_ECPU_BIT_NUM) & 1;
}
static bool mlx5_ecpf_esw_admins_host_pf(const struct mlx5_core_dev *dev)
{
/* In separate host mode, PF enables itself.
* When ECPF is eswitch manager, eswitch enables host PF after
* eswitch is setup.
*/
return mlx5_core_is_ecpf_esw_manager(dev);
}
int mlx5_cmd_host_pf_enable_hca(struct mlx5_core_dev *dev)
{
u32 out[MLX5_ST_SZ_DW(enable_hca_out)] = {};
u32 in[MLX5_ST_SZ_DW(enable_hca_in)] = {};
MLX5_SET(enable_hca_in, in, opcode, MLX5_CMD_OP_ENABLE_HCA);
MLX5_SET(enable_hca_in, in, function_id, 0);
MLX5_SET(enable_hca_in, in, embedded_cpu_function, 0);
return mlx5_cmd_exec(dev, &in, sizeof(in), &out, sizeof(out));
}
int mlx5_cmd_host_pf_disable_hca(struct mlx5_core_dev *dev)
{
u32 out[MLX5_ST_SZ_DW(disable_hca_out)] = {};
u32 in[MLX5_ST_SZ_DW(disable_hca_in)] = {};
MLX5_SET(disable_hca_in, in, opcode, MLX5_CMD_OP_DISABLE_HCA);
MLX5_SET(disable_hca_in, in, function_id, 0);
MLX5_SET(disable_hca_in, in, embedded_cpu_function, 0);
return mlx5_cmd_exec(dev, in, sizeof(in), out, sizeof(out));
}
static int mlx5_host_pf_init(struct mlx5_core_dev *dev)
{
int err;
if (mlx5_ecpf_esw_admins_host_pf(dev))
return 0;
/* ECPF shall enable HCA for host PF in the same way a PF
* does this for its VFs when ECPF is not a eswitch manager.
*/
err = mlx5_cmd_host_pf_enable_hca(dev);
if (err)
mlx5_core_err(dev, "Failed to enable external host PF HCA err(%d)\n", err);
return err;
}
static void mlx5_host_pf_cleanup(struct mlx5_core_dev *dev)
{
int err;
if (mlx5_ecpf_esw_admins_host_pf(dev))
return;
err = mlx5_cmd_host_pf_disable_hca(dev);
if (err) {
mlx5_core_err(dev, "Failed to disable external host PF HCA err(%d)\n", err);
return;
}
}
int mlx5_ec_init(struct mlx5_core_dev *dev)
{
if (!mlx5_core_is_ecpf(dev))
return 0;
return mlx5_host_pf_init(dev);
}
void mlx5_ec_cleanup(struct mlx5_core_dev *dev)
{
int err;
if (!mlx5_core_is_ecpf(dev))
return;
mlx5_host_pf_cleanup(dev);
err = mlx5_wait_for_pages(dev, &dev->priv.page_counters[MLX5_HOST_PF]);
if (err)
mlx5_core_warn(dev, "Timeout reclaiming external host PF pages err(%d)\n", err);
err = mlx5_wait_for_pages(dev, &dev->priv.page_counters[MLX5_VF]);
if (err)
mlx5_core_warn(dev, "Timeout reclaiming external host VFs pages err(%d)\n", err);
}