diff options
Diffstat (limited to 'drivers/net/ethernet/mellanox/mlx5/core/main.c')
| -rw-r--r-- | drivers/net/ethernet/mellanox/mlx5/core/main.c | 548 | 
1 files changed, 548 insertions, 0 deletions
diff --git a/drivers/net/ethernet/mellanox/mlx5/core/main.c b/drivers/net/ethernet/mellanox/mlx5/core/main.c new file mode 100644 index 00000000000..ee24f132e31 --- /dev/null +++ b/drivers/net/ethernet/mellanox/mlx5/core/main.c @@ -0,0 +1,548 @@ +/* + * Copyright (c) 2013, Mellanox Technologies inc.  All rights reserved. + * + * This software is available to you under a choice of one of two + * licenses.  You may choose to be licensed under the terms of the GNU + * General Public License (GPL) Version 2, available from the file + * COPYING in the main directory of this source tree, or the + * OpenIB.org BSD license below: + * + *     Redistribution and use in source and binary forms, with or + *     without modification, are permitted provided that the following + *     conditions are met: + * + *      - Redistributions of source code must retain the above + *        copyright notice, this list of conditions and the following + *        disclaimer. + * + *      - Redistributions in binary form must reproduce the above + *        copyright notice, this list of conditions and the following + *        disclaimer in the documentation and/or other materials + *        provided with the distribution. + * + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, + * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF + * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND + * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS + * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN + * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN + * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE + * SOFTWARE. + */ + +#include <asm-generic/kmap_types.h> +#include <linux/module.h> +#include <linux/init.h> +#include <linux/errno.h> +#include <linux/pci.h> +#include <linux/dma-mapping.h> +#include <linux/slab.h> +#include <linux/io-mapping.h> +#include <linux/mlx5/driver.h> +#include <linux/mlx5/cq.h> +#include <linux/mlx5/qp.h> +#include <linux/mlx5/srq.h> +#include <linux/debugfs.h> +#include "mlx5_core.h" + +#define DRIVER_NAME "mlx5_core" +#define DRIVER_VERSION "2.2-1" +#define DRIVER_RELDATE	"Feb 2014" + +MODULE_AUTHOR("Eli Cohen <eli@mellanox.com>"); +MODULE_DESCRIPTION("Mellanox ConnectX-IB HCA core library"); +MODULE_LICENSE("Dual BSD/GPL"); +MODULE_VERSION(DRIVER_VERSION); + +int mlx5_core_debug_mask; +module_param_named(debug_mask, mlx5_core_debug_mask, int, 0644); +MODULE_PARM_DESC(debug_mask, "debug mask: 1 = dump cmd data, 2 = dump cmd exec time, 3 = both. Default=0"); + +struct workqueue_struct *mlx5_core_wq; + +static int set_dma_caps(struct pci_dev *pdev) +{ +	int err; + +	err = pci_set_dma_mask(pdev, DMA_BIT_MASK(64)); +	if (err) { +		dev_warn(&pdev->dev, "Warning: couldn't set 64-bit PCI DMA mask\n"); +		err = pci_set_dma_mask(pdev, DMA_BIT_MASK(32)); +		if (err) { +			dev_err(&pdev->dev, "Can't set PCI DMA mask, aborting\n"); +			return err; +		} +	} + +	err = pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(64)); +	if (err) { +		dev_warn(&pdev->dev, +			 "Warning: couldn't set 64-bit consistent PCI DMA mask\n"); +		err = pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(32)); +		if (err) { +			dev_err(&pdev->dev, +				"Can't set consistent PCI DMA mask, aborting\n"); +			return err; +		} +	} + +	dma_set_max_seg_size(&pdev->dev, 2u * 1024 * 1024 * 1024); +	return err; +} + +static int request_bar(struct pci_dev *pdev) +{ +	int err = 0; + +	if (!(pci_resource_flags(pdev, 0) & IORESOURCE_MEM)) { +		dev_err(&pdev->dev, "Missing registers BAR, aborting\n"); +		return -ENODEV; +	} + +	err = pci_request_regions(pdev, DRIVER_NAME); +	if (err) +		dev_err(&pdev->dev, "Couldn't get PCI resources, aborting\n"); + +	return err; +} + +static void release_bar(struct pci_dev *pdev) +{ +	pci_release_regions(pdev); +} + +static int mlx5_enable_msix(struct mlx5_core_dev *dev) +{ +	struct mlx5_eq_table *table = &dev->priv.eq_table; +	int num_eqs = 1 << dev->caps.log_max_eq; +	int nvec; +	int i; + +	nvec = dev->caps.num_ports * num_online_cpus() + MLX5_EQ_VEC_COMP_BASE; +	nvec = min_t(int, nvec, num_eqs); +	if (nvec <= MLX5_EQ_VEC_COMP_BASE) +		return -ENOMEM; + +	table->msix_arr = kzalloc(nvec * sizeof(*table->msix_arr), GFP_KERNEL); +	if (!table->msix_arr) +		return -ENOMEM; + +	for (i = 0; i < nvec; i++) +		table->msix_arr[i].entry = i; + +	nvec = pci_enable_msix_range(dev->pdev, table->msix_arr, +				     MLX5_EQ_VEC_COMP_BASE, nvec); +	if (nvec < 0) +		return nvec; + +	table->num_comp_vectors = nvec - MLX5_EQ_VEC_COMP_BASE; + +	return 0; +} + +static void mlx5_disable_msix(struct mlx5_core_dev *dev) +{ +	struct mlx5_eq_table *table = &dev->priv.eq_table; + +	pci_disable_msix(dev->pdev); +	kfree(table->msix_arr); +} + +struct mlx5_reg_host_endianess { +	u8	he; +	u8      rsvd[15]; +}; + + +#define CAP_MASK(pos, size) ((u64)((1 << (size)) - 1) << (pos)) + +enum { +	MLX5_CAP_BITS_RW_MASK	= CAP_MASK(MLX5_CAP_OFF_CMDIF_CSUM, 2) | +				  CAP_MASK(MLX5_CAP_OFF_DCT, 1), +}; + +/* selectively copy writable fields clearing any reserved area + */ +static void copy_rw_fields(struct mlx5_hca_cap *to, struct mlx5_hca_cap *from) +{ +	u64 v64; + +	to->log_max_qp = from->log_max_qp & 0x1f; +	to->log_max_ra_req_dc = from->log_max_ra_req_dc & 0x3f; +	to->log_max_ra_res_dc = from->log_max_ra_res_dc & 0x3f; +	to->log_max_ra_req_qp = from->log_max_ra_req_qp & 0x3f; +	to->log_max_ra_res_qp = from->log_max_ra_res_qp & 0x3f; +	to->log_max_atomic_size_qp = from->log_max_atomic_size_qp; +	to->log_max_atomic_size_dc = from->log_max_atomic_size_dc; +	v64 = be64_to_cpu(from->flags) & MLX5_CAP_BITS_RW_MASK; +	to->flags = cpu_to_be64(v64); +} + +enum { +	HCA_CAP_OPMOD_GET_MAX	= 0, +	HCA_CAP_OPMOD_GET_CUR	= 1, +}; + +static int handle_hca_cap(struct mlx5_core_dev *dev) +{ +	struct mlx5_cmd_query_hca_cap_mbox_out *query_out = NULL; +	struct mlx5_cmd_set_hca_cap_mbox_in *set_ctx = NULL; +	struct mlx5_cmd_query_hca_cap_mbox_in query_ctx; +	struct mlx5_cmd_set_hca_cap_mbox_out set_out; +	u64 flags; +	int err; + +	memset(&query_ctx, 0, sizeof(query_ctx)); +	query_out = kzalloc(sizeof(*query_out), GFP_KERNEL); +	if (!query_out) +		return -ENOMEM; + +	set_ctx = kzalloc(sizeof(*set_ctx), GFP_KERNEL); +	if (!set_ctx) { +		err = -ENOMEM; +		goto query_ex; +	} + +	query_ctx.hdr.opcode = cpu_to_be16(MLX5_CMD_OP_QUERY_HCA_CAP); +	query_ctx.hdr.opmod  = cpu_to_be16(HCA_CAP_OPMOD_GET_CUR); +	err = mlx5_cmd_exec(dev, &query_ctx, sizeof(query_ctx), +				 query_out, sizeof(*query_out)); +	if (err) +		goto query_ex; + +	err = mlx5_cmd_status_to_err(&query_out->hdr); +	if (err) { +		mlx5_core_warn(dev, "query hca cap failed, %d\n", err); +		goto query_ex; +	} + +	copy_rw_fields(&set_ctx->hca_cap, &query_out->hca_cap); + +	if (dev->profile->mask & MLX5_PROF_MASK_QP_SIZE) +		set_ctx->hca_cap.log_max_qp = dev->profile->log_max_qp; + +	flags = be64_to_cpu(query_out->hca_cap.flags); +	/* disable checksum */ +	flags &= ~MLX5_DEV_CAP_FLAG_CMDIF_CSUM; + +	set_ctx->hca_cap.flags = cpu_to_be64(flags); +	memset(&set_out, 0, sizeof(set_out)); +	set_ctx->hca_cap.log_uar_page_sz = cpu_to_be16(PAGE_SHIFT - 12); +	set_ctx->hdr.opcode = cpu_to_be16(MLX5_CMD_OP_SET_HCA_CAP); +	err = mlx5_cmd_exec(dev, set_ctx, sizeof(*set_ctx), +				 &set_out, sizeof(set_out)); +	if (err) { +		mlx5_core_warn(dev, "set hca cap failed, %d\n", err); +		goto query_ex; +	} + +	err = mlx5_cmd_status_to_err(&set_out.hdr); +	if (err) +		goto query_ex; + +query_ex: +	kfree(query_out); +	kfree(set_ctx); + +	return err; +} + +static int set_hca_ctrl(struct mlx5_core_dev *dev) +{ +	struct mlx5_reg_host_endianess he_in; +	struct mlx5_reg_host_endianess he_out; +	int err; + +	memset(&he_in, 0, sizeof(he_in)); +	he_in.he = MLX5_SET_HOST_ENDIANNESS; +	err = mlx5_core_access_reg(dev, &he_in,  sizeof(he_in), +					&he_out, sizeof(he_out), +					MLX5_REG_HOST_ENDIANNESS, 0, 1); +	return err; +} + +static int mlx5_core_enable_hca(struct mlx5_core_dev *dev) +{ +	int err; +	struct mlx5_enable_hca_mbox_in in; +	struct mlx5_enable_hca_mbox_out out; + +	memset(&in, 0, sizeof(in)); +	memset(&out, 0, sizeof(out)); +	in.hdr.opcode = cpu_to_be16(MLX5_CMD_OP_ENABLE_HCA); +	err = mlx5_cmd_exec(dev, &in, sizeof(in), &out, sizeof(out)); +	if (err) +		return err; + +	if (out.hdr.status) +		return mlx5_cmd_status_to_err(&out.hdr); + +	return 0; +} + +static int mlx5_core_disable_hca(struct mlx5_core_dev *dev) +{ +	int err; +	struct mlx5_disable_hca_mbox_in in; +	struct mlx5_disable_hca_mbox_out out; + +	memset(&in, 0, sizeof(in)); +	memset(&out, 0, sizeof(out)); +	in.hdr.opcode = cpu_to_be16(MLX5_CMD_OP_DISABLE_HCA); +	err = mlx5_cmd_exec(dev, &in, sizeof(in), &out, sizeof(out)); +	if (err) +		return err; + +	if (out.hdr.status) +		return mlx5_cmd_status_to_err(&out.hdr); + +	return 0; +} + +int mlx5_dev_init(struct mlx5_core_dev *dev, struct pci_dev *pdev) +{ +	struct mlx5_priv *priv = &dev->priv; +	int err; + +	dev->pdev = pdev; +	pci_set_drvdata(dev->pdev, dev); +	strncpy(priv->name, dev_name(&pdev->dev), MLX5_MAX_NAME_LEN); +	priv->name[MLX5_MAX_NAME_LEN - 1] = 0; + +	mutex_init(&priv->pgdir_mutex); +	INIT_LIST_HEAD(&priv->pgdir_list); +	spin_lock_init(&priv->mkey_lock); + +	priv->dbg_root = debugfs_create_dir(dev_name(&pdev->dev), mlx5_debugfs_root); +	if (!priv->dbg_root) +		return -ENOMEM; + +	err = pci_enable_device(pdev); +	if (err) { +		dev_err(&pdev->dev, "Cannot enable PCI device, aborting\n"); +		goto err_dbg; +	} + +	err = request_bar(pdev); +	if (err) { +		dev_err(&pdev->dev, "error requesting BARs, aborting\n"); +		goto err_disable; +	} + +	pci_set_master(pdev); + +	err = set_dma_caps(pdev); +	if (err) { +		dev_err(&pdev->dev, "Failed setting DMA capabilities mask, aborting\n"); +		goto err_clr_master; +	} + +	dev->iseg_base = pci_resource_start(dev->pdev, 0); +	dev->iseg = ioremap(dev->iseg_base, sizeof(*dev->iseg)); +	if (!dev->iseg) { +		err = -ENOMEM; +		dev_err(&pdev->dev, "Failed mapping initialization segment, aborting\n"); +		goto err_clr_master; +	} +	dev_info(&pdev->dev, "firmware version: %d.%d.%d\n", fw_rev_maj(dev), +		 fw_rev_min(dev), fw_rev_sub(dev)); + +	err = mlx5_cmd_init(dev); +	if (err) { +		dev_err(&pdev->dev, "Failed initializing command interface, aborting\n"); +		goto err_unmap; +	} + +	mlx5_pagealloc_init(dev); + +	err = mlx5_core_enable_hca(dev); +	if (err) { +		dev_err(&pdev->dev, "enable hca failed\n"); +		goto err_pagealloc_cleanup; +	} + +	err = mlx5_satisfy_startup_pages(dev, 1); +	if (err) { +		dev_err(&pdev->dev, "failed to allocate boot pages\n"); +		goto err_disable_hca; +	} + +	err = set_hca_ctrl(dev); +	if (err) { +		dev_err(&pdev->dev, "set_hca_ctrl failed\n"); +		goto reclaim_boot_pages; +	} + +	err = handle_hca_cap(dev); +	if (err) { +		dev_err(&pdev->dev, "handle_hca_cap failed\n"); +		goto reclaim_boot_pages; +	} + +	err = mlx5_satisfy_startup_pages(dev, 0); +	if (err) { +		dev_err(&pdev->dev, "failed to allocate init pages\n"); +		goto reclaim_boot_pages; +	} + +	err = mlx5_pagealloc_start(dev); +	if (err) { +		dev_err(&pdev->dev, "mlx5_pagealloc_start failed\n"); +		goto reclaim_boot_pages; +	} + +	err = mlx5_cmd_init_hca(dev); +	if (err) { +		dev_err(&pdev->dev, "init hca failed\n"); +		goto err_pagealloc_stop; +	} + +	mlx5_start_health_poll(dev); + +	err = mlx5_cmd_query_hca_cap(dev, &dev->caps); +	if (err) { +		dev_err(&pdev->dev, "query hca failed\n"); +		goto err_stop_poll; +	} + +	err = mlx5_cmd_query_adapter(dev); +	if (err) { +		dev_err(&pdev->dev, "query adapter failed\n"); +		goto err_stop_poll; +	} + +	err = mlx5_enable_msix(dev); +	if (err) { +		dev_err(&pdev->dev, "enable msix failed\n"); +		goto err_stop_poll; +	} + +	err = mlx5_eq_init(dev); +	if (err) { +		dev_err(&pdev->dev, "failed to initialize eq\n"); +		goto disable_msix; +	} + +	err = mlx5_alloc_uuars(dev, &priv->uuari); +	if (err) { +		dev_err(&pdev->dev, "Failed allocating uar, aborting\n"); +		goto err_eq_cleanup; +	} + +	err = mlx5_start_eqs(dev); +	if (err) { +		dev_err(&pdev->dev, "Failed to start pages and async EQs\n"); +		goto err_free_uar; +	} + +	MLX5_INIT_DOORBELL_LOCK(&priv->cq_uar_lock); + +	mlx5_init_cq_table(dev); +	mlx5_init_qp_table(dev); +	mlx5_init_srq_table(dev); +	mlx5_init_mr_table(dev); + +	return 0; + +err_free_uar: +	mlx5_free_uuars(dev, &priv->uuari); + +err_eq_cleanup: +	mlx5_eq_cleanup(dev); + +disable_msix: +	mlx5_disable_msix(dev); + +err_stop_poll: +	mlx5_stop_health_poll(dev); +	if (mlx5_cmd_teardown_hca(dev)) { +		dev_err(&dev->pdev->dev, "tear_down_hca failed, skip cleanup\n"); +		return err; +	} + +err_pagealloc_stop: +	mlx5_pagealloc_stop(dev); + +reclaim_boot_pages: +	mlx5_reclaim_startup_pages(dev); + +err_disable_hca: +	mlx5_core_disable_hca(dev); + +err_pagealloc_cleanup: +	mlx5_pagealloc_cleanup(dev); +	mlx5_cmd_cleanup(dev); + +err_unmap: +	iounmap(dev->iseg); + +err_clr_master: +	pci_clear_master(dev->pdev); +	release_bar(dev->pdev); + +err_disable: +	pci_disable_device(dev->pdev); + +err_dbg: +	debugfs_remove(priv->dbg_root); +	return err; +} +EXPORT_SYMBOL(mlx5_dev_init); + +void mlx5_dev_cleanup(struct mlx5_core_dev *dev) +{ +	struct mlx5_priv *priv = &dev->priv; + +	mlx5_cleanup_srq_table(dev); +	mlx5_cleanup_qp_table(dev); +	mlx5_cleanup_cq_table(dev); +	mlx5_stop_eqs(dev); +	mlx5_free_uuars(dev, &priv->uuari); +	mlx5_eq_cleanup(dev); +	mlx5_disable_msix(dev); +	mlx5_stop_health_poll(dev); +	if (mlx5_cmd_teardown_hca(dev)) { +		dev_err(&dev->pdev->dev, "tear_down_hca failed, skip cleanup\n"); +		return; +	} +	mlx5_pagealloc_stop(dev); +	mlx5_reclaim_startup_pages(dev); +	mlx5_core_disable_hca(dev); +	mlx5_pagealloc_cleanup(dev); +	mlx5_cmd_cleanup(dev); +	iounmap(dev->iseg); +	pci_clear_master(dev->pdev); +	release_bar(dev->pdev); +	pci_disable_device(dev->pdev); +	debugfs_remove(priv->dbg_root); +} +EXPORT_SYMBOL(mlx5_dev_cleanup); + +static int __init init(void) +{ +	int err; + +	mlx5_register_debugfs(); +	mlx5_core_wq = create_singlethread_workqueue("mlx5_core_wq"); +	if (!mlx5_core_wq) { +		err = -ENOMEM; +		goto err_debug; +	} +	mlx5_health_init(); + +	return 0; + +err_debug: +	mlx5_unregister_debugfs(); +	return err; +} + +static void __exit cleanup(void) +{ +	mlx5_health_cleanup(); +	destroy_workqueue(mlx5_core_wq); +	mlx5_unregister_debugfs(); +} + +module_init(init); +module_exit(cleanup);  | 
