5a0e3ad6af
percpu.h is included by sched.h and module.h and thus ends up being included when building most .c files. percpu.h includes slab.h which in turn includes gfp.h making everything defined by the two files universally available and complicating inclusion dependencies. percpu.h -> slab.h dependency is about to be removed. Prepare for this change by updating users of gfp and slab facilities include those headers directly instead of assuming availability. As this conversion needs to touch large number of source files, the following script is used as the basis of conversion. http://userweb.kernel.org/~tj/misc/slabh-sweep.py The script does the followings. * Scan files for gfp and slab usages and update includes such that only the necessary includes are there. ie. if only gfp is used, gfp.h, if slab is used, slab.h. * When the script inserts a new include, it looks at the include blocks and try to put the new include such that its order conforms to its surrounding. It's put in the include block which contains core kernel includes, in the same order that the rest are ordered - alphabetical, Christmas tree, rev-Xmas-tree or at the end if there doesn't seem to be any matching order. * If the script can't find a place to put a new include (mostly because the file doesn't have fitting include block), it prints out an error message indicating which .h file needs to be added to the file. The conversion was done in the following steps. 1. The initial automatic conversion of all .c files updated slightly over 4000 files, deleting around 700 includes and adding ~480 gfp.h and ~3000 slab.h inclusions. The script emitted errors for ~400 files. 2. Each error was manually checked. Some didn't need the inclusion, some needed manual addition while adding it to implementation .h or embedding .c file was more appropriate for others. This step added inclusions to around 150 files. 3. The script was run again and the output was compared to the edits from #2 to make sure no file was left behind. 4. Several build tests were done and a couple of problems were fixed. e.g. lib/decompress_*.c used malloc/free() wrappers around slab APIs requiring slab.h to be added manually. 5. The script was run on all .h files but without automatically editing them as sprinkling gfp.h and slab.h inclusions around .h files could easily lead to inclusion dependency hell. Most gfp.h inclusion directives were ignored as stuff from gfp.h was usually wildly available and often used in preprocessor macros. Each slab.h inclusion directive was examined and added manually as necessary. 6. percpu.h was updated not to include slab.h. 7. Build test were done on the following configurations and failures were fixed. CONFIG_GCOV_KERNEL was turned off for all tests (as my distributed build env didn't work with gcov compiles) and a few more options had to be turned off depending on archs to make things build (like ipr on powerpc/64 which failed due to missing writeq). * x86 and x86_64 UP and SMP allmodconfig and a custom test config. * powerpc and powerpc64 SMP allmodconfig * sparc and sparc64 SMP allmodconfig * ia64 SMP allmodconfig * s390 SMP allmodconfig * alpha SMP allmodconfig * um on x86_64 SMP allmodconfig 8. percpu.h modifications were reverted so that it could be applied as a separate patch and serve as bisection point. Given the fact that I had only a couple of failures from tests on step 6, I'm fairly confident about the coverage of this conversion patch. If there is a breakage, it's likely to be something in one of the arch headers which should be easily discoverable easily on most builds of the specific arch. Signed-off-by: Tejun Heo <tj@kernel.org> Guess-its-ok-by: Christoph Lameter <cl@linux-foundation.org> Cc: Ingo Molnar <mingo@redhat.com> Cc: Lee Schermerhorn <Lee.Schermerhorn@hp.com>
1357 lines
36 KiB
C
1357 lines
36 KiB
C
/*
|
|
* Copyright (c) 2004, 2005 Topspin Communications. All rights reserved.
|
|
* Copyright (c) 2005 Sun Microsystems, Inc. All rights reserved.
|
|
* Copyright (c) 2005 Mellanox Technologies. All rights reserved.
|
|
*
|
|
* This software is available to you under a choice of one of two
|
|
* licenses. You may choose to be licensed under the terms of the GNU
|
|
* General Public License (GPL) Version 2, available from the file
|
|
* COPYING in the main directory of this source tree, or the
|
|
* OpenIB.org BSD license below:
|
|
*
|
|
* Redistribution and use in source and binary forms, with or
|
|
* without modification, are permitted provided that the following
|
|
* conditions are met:
|
|
*
|
|
* - Redistributions of source code must retain the above
|
|
* copyright notice, this list of conditions and the following
|
|
* disclaimer.
|
|
*
|
|
* - Redistributions in binary form must reproduce the above
|
|
* copyright notice, this list of conditions and the following
|
|
* disclaimer in the documentation and/or other materials
|
|
* provided with the distribution.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
|
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
|
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
|
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
|
|
* BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
|
|
* ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
|
|
* CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
|
* SOFTWARE.
|
|
*/
|
|
|
|
#include <linux/module.h>
|
|
#include <linux/init.h>
|
|
#include <linux/errno.h>
|
|
#include <linux/pci.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/gfp.h>
|
|
|
|
#include "mthca_dev.h"
|
|
#include "mthca_config_reg.h"
|
|
#include "mthca_cmd.h"
|
|
#include "mthca_profile.h"
|
|
#include "mthca_memfree.h"
|
|
#include "mthca_wqe.h"
|
|
|
|
MODULE_AUTHOR("Roland Dreier");
|
|
MODULE_DESCRIPTION("Mellanox InfiniBand HCA low-level driver");
|
|
MODULE_LICENSE("Dual BSD/GPL");
|
|
MODULE_VERSION(DRV_VERSION);
|
|
|
|
#ifdef CONFIG_INFINIBAND_MTHCA_DEBUG
|
|
|
|
int mthca_debug_level = 0;
|
|
module_param_named(debug_level, mthca_debug_level, int, 0644);
|
|
MODULE_PARM_DESC(debug_level, "Enable debug tracing if > 0");
|
|
|
|
#endif /* CONFIG_INFINIBAND_MTHCA_DEBUG */
|
|
|
|
#ifdef CONFIG_PCI_MSI
|
|
|
|
static int msi_x = 1;
|
|
module_param(msi_x, int, 0444);
|
|
MODULE_PARM_DESC(msi_x, "attempt to use MSI-X if nonzero");
|
|
|
|
#else /* CONFIG_PCI_MSI */
|
|
|
|
#define msi_x (0)
|
|
|
|
#endif /* CONFIG_PCI_MSI */
|
|
|
|
static int tune_pci = 0;
|
|
module_param(tune_pci, int, 0444);
|
|
MODULE_PARM_DESC(tune_pci, "increase PCI burst from the default set by BIOS if nonzero");
|
|
|
|
DEFINE_MUTEX(mthca_device_mutex);
|
|
|
|
#define MTHCA_DEFAULT_NUM_QP (1 << 16)
|
|
#define MTHCA_DEFAULT_RDB_PER_QP (1 << 2)
|
|
#define MTHCA_DEFAULT_NUM_CQ (1 << 16)
|
|
#define MTHCA_DEFAULT_NUM_MCG (1 << 13)
|
|
#define MTHCA_DEFAULT_NUM_MPT (1 << 17)
|
|
#define MTHCA_DEFAULT_NUM_MTT (1 << 20)
|
|
#define MTHCA_DEFAULT_NUM_UDAV (1 << 15)
|
|
#define MTHCA_DEFAULT_NUM_RESERVED_MTTS (1 << 18)
|
|
#define MTHCA_DEFAULT_NUM_UARC_SIZE (1 << 18)
|
|
|
|
static struct mthca_profile hca_profile = {
|
|
.num_qp = MTHCA_DEFAULT_NUM_QP,
|
|
.rdb_per_qp = MTHCA_DEFAULT_RDB_PER_QP,
|
|
.num_cq = MTHCA_DEFAULT_NUM_CQ,
|
|
.num_mcg = MTHCA_DEFAULT_NUM_MCG,
|
|
.num_mpt = MTHCA_DEFAULT_NUM_MPT,
|
|
.num_mtt = MTHCA_DEFAULT_NUM_MTT,
|
|
.num_udav = MTHCA_DEFAULT_NUM_UDAV, /* Tavor only */
|
|
.fmr_reserved_mtts = MTHCA_DEFAULT_NUM_RESERVED_MTTS, /* Tavor only */
|
|
.uarc_size = MTHCA_DEFAULT_NUM_UARC_SIZE, /* Arbel only */
|
|
};
|
|
|
|
module_param_named(num_qp, hca_profile.num_qp, int, 0444);
|
|
MODULE_PARM_DESC(num_qp, "maximum number of QPs per HCA");
|
|
|
|
module_param_named(rdb_per_qp, hca_profile.rdb_per_qp, int, 0444);
|
|
MODULE_PARM_DESC(rdb_per_qp, "number of RDB buffers per QP");
|
|
|
|
module_param_named(num_cq, hca_profile.num_cq, int, 0444);
|
|
MODULE_PARM_DESC(num_cq, "maximum number of CQs per HCA");
|
|
|
|
module_param_named(num_mcg, hca_profile.num_mcg, int, 0444);
|
|
MODULE_PARM_DESC(num_mcg, "maximum number of multicast groups per HCA");
|
|
|
|
module_param_named(num_mpt, hca_profile.num_mpt, int, 0444);
|
|
MODULE_PARM_DESC(num_mpt,
|
|
"maximum number of memory protection table entries per HCA");
|
|
|
|
module_param_named(num_mtt, hca_profile.num_mtt, int, 0444);
|
|
MODULE_PARM_DESC(num_mtt,
|
|
"maximum number of memory translation table segments per HCA");
|
|
|
|
module_param_named(num_udav, hca_profile.num_udav, int, 0444);
|
|
MODULE_PARM_DESC(num_udav, "maximum number of UD address vectors per HCA");
|
|
|
|
module_param_named(fmr_reserved_mtts, hca_profile.fmr_reserved_mtts, int, 0444);
|
|
MODULE_PARM_DESC(fmr_reserved_mtts,
|
|
"number of memory translation table segments reserved for FMR");
|
|
|
|
static int log_mtts_per_seg = ilog2(MTHCA_MTT_SEG_SIZE / 8);
|
|
module_param_named(log_mtts_per_seg, log_mtts_per_seg, int, 0444);
|
|
MODULE_PARM_DESC(log_mtts_per_seg, "Log2 number of MTT entries per segment (1-5)");
|
|
|
|
static char mthca_version[] __devinitdata =
|
|
DRV_NAME ": Mellanox InfiniBand HCA driver v"
|
|
DRV_VERSION " (" DRV_RELDATE ")\n";
|
|
|
|
static int mthca_tune_pci(struct mthca_dev *mdev)
|
|
{
|
|
if (!tune_pci)
|
|
return 0;
|
|
|
|
/* First try to max out Read Byte Count */
|
|
if (pci_find_capability(mdev->pdev, PCI_CAP_ID_PCIX)) {
|
|
if (pcix_set_mmrbc(mdev->pdev, pcix_get_max_mmrbc(mdev->pdev))) {
|
|
mthca_err(mdev, "Couldn't set PCI-X max read count, "
|
|
"aborting.\n");
|
|
return -ENODEV;
|
|
}
|
|
} else if (!(mdev->mthca_flags & MTHCA_FLAG_PCIE))
|
|
mthca_info(mdev, "No PCI-X capability, not setting RBC.\n");
|
|
|
|
if (pci_find_capability(mdev->pdev, PCI_CAP_ID_EXP)) {
|
|
if (pcie_set_readrq(mdev->pdev, 4096)) {
|
|
mthca_err(mdev, "Couldn't write PCI Express read request, "
|
|
"aborting.\n");
|
|
return -ENODEV;
|
|
}
|
|
} else if (mdev->mthca_flags & MTHCA_FLAG_PCIE)
|
|
mthca_info(mdev, "No PCI Express capability, "
|
|
"not setting Max Read Request Size.\n");
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int mthca_dev_lim(struct mthca_dev *mdev, struct mthca_dev_lim *dev_lim)
|
|
{
|
|
int err;
|
|
u8 status;
|
|
|
|
mdev->limits.mtt_seg_size = (1 << log_mtts_per_seg) * 8;
|
|
err = mthca_QUERY_DEV_LIM(mdev, dev_lim, &status);
|
|
if (err) {
|
|
mthca_err(mdev, "QUERY_DEV_LIM command failed, aborting.\n");
|
|
return err;
|
|
}
|
|
if (status) {
|
|
mthca_err(mdev, "QUERY_DEV_LIM returned status 0x%02x, "
|
|
"aborting.\n", status);
|
|
return -EINVAL;
|
|
}
|
|
if (dev_lim->min_page_sz > PAGE_SIZE) {
|
|
mthca_err(mdev, "HCA minimum page size of %d bigger than "
|
|
"kernel PAGE_SIZE of %ld, aborting.\n",
|
|
dev_lim->min_page_sz, PAGE_SIZE);
|
|
return -ENODEV;
|
|
}
|
|
if (dev_lim->num_ports > MTHCA_MAX_PORTS) {
|
|
mthca_err(mdev, "HCA has %d ports, but we only support %d, "
|
|
"aborting.\n",
|
|
dev_lim->num_ports, MTHCA_MAX_PORTS);
|
|
return -ENODEV;
|
|
}
|
|
|
|
if (dev_lim->uar_size > pci_resource_len(mdev->pdev, 2)) {
|
|
mthca_err(mdev, "HCA reported UAR size of 0x%x bigger than "
|
|
"PCI resource 2 size of 0x%llx, aborting.\n",
|
|
dev_lim->uar_size,
|
|
(unsigned long long)pci_resource_len(mdev->pdev, 2));
|
|
return -ENODEV;
|
|
}
|
|
|
|
mdev->limits.num_ports = dev_lim->num_ports;
|
|
mdev->limits.vl_cap = dev_lim->max_vl;
|
|
mdev->limits.mtu_cap = dev_lim->max_mtu;
|
|
mdev->limits.gid_table_len = dev_lim->max_gids;
|
|
mdev->limits.pkey_table_len = dev_lim->max_pkeys;
|
|
mdev->limits.local_ca_ack_delay = dev_lim->local_ca_ack_delay;
|
|
/*
|
|
* Need to allow for worst case send WQE overhead and check
|
|
* whether max_desc_sz imposes a lower limit than max_sg; UD
|
|
* send has the biggest overhead.
|
|
*/
|
|
mdev->limits.max_sg = min_t(int, dev_lim->max_sg,
|
|
(dev_lim->max_desc_sz -
|
|
sizeof (struct mthca_next_seg) -
|
|
(mthca_is_memfree(mdev) ?
|
|
sizeof (struct mthca_arbel_ud_seg) :
|
|
sizeof (struct mthca_tavor_ud_seg))) /
|
|
sizeof (struct mthca_data_seg));
|
|
mdev->limits.max_wqes = dev_lim->max_qp_sz;
|
|
mdev->limits.max_qp_init_rdma = dev_lim->max_requester_per_qp;
|
|
mdev->limits.reserved_qps = dev_lim->reserved_qps;
|
|
mdev->limits.max_srq_wqes = dev_lim->max_srq_sz;
|
|
mdev->limits.reserved_srqs = dev_lim->reserved_srqs;
|
|
mdev->limits.reserved_eecs = dev_lim->reserved_eecs;
|
|
mdev->limits.max_desc_sz = dev_lim->max_desc_sz;
|
|
mdev->limits.max_srq_sge = mthca_max_srq_sge(mdev);
|
|
/*
|
|
* Subtract 1 from the limit because we need to allocate a
|
|
* spare CQE so the HCA HW can tell the difference between an
|
|
* empty CQ and a full CQ.
|
|
*/
|
|
mdev->limits.max_cqes = dev_lim->max_cq_sz - 1;
|
|
mdev->limits.reserved_cqs = dev_lim->reserved_cqs;
|
|
mdev->limits.reserved_eqs = dev_lim->reserved_eqs;
|
|
mdev->limits.reserved_mtts = dev_lim->reserved_mtts;
|
|
mdev->limits.reserved_mrws = dev_lim->reserved_mrws;
|
|
mdev->limits.reserved_uars = dev_lim->reserved_uars;
|
|
mdev->limits.reserved_pds = dev_lim->reserved_pds;
|
|
mdev->limits.port_width_cap = dev_lim->max_port_width;
|
|
mdev->limits.page_size_cap = ~(u32) (dev_lim->min_page_sz - 1);
|
|
mdev->limits.flags = dev_lim->flags;
|
|
/*
|
|
* For old FW that doesn't return static rate support, use a
|
|
* value of 0x3 (only static rate values of 0 or 1 are handled),
|
|
* except on Sinai, where even old FW can handle static rate
|
|
* values of 2 and 3.
|
|
*/
|
|
if (dev_lim->stat_rate_support)
|
|
mdev->limits.stat_rate_support = dev_lim->stat_rate_support;
|
|
else if (mdev->mthca_flags & MTHCA_FLAG_SINAI_OPT)
|
|
mdev->limits.stat_rate_support = 0xf;
|
|
else
|
|
mdev->limits.stat_rate_support = 0x3;
|
|
|
|
/* IB_DEVICE_RESIZE_MAX_WR not supported by driver.
|
|
May be doable since hardware supports it for SRQ.
|
|
|
|
IB_DEVICE_N_NOTIFY_CQ is supported by hardware but not by driver.
|
|
|
|
IB_DEVICE_SRQ_RESIZE is supported by hardware but SRQ is not
|
|
supported by driver. */
|
|
mdev->device_cap_flags = IB_DEVICE_CHANGE_PHY_PORT |
|
|
IB_DEVICE_PORT_ACTIVE_EVENT |
|
|
IB_DEVICE_SYS_IMAGE_GUID |
|
|
IB_DEVICE_RC_RNR_NAK_GEN;
|
|
|
|
if (dev_lim->flags & DEV_LIM_FLAG_BAD_PKEY_CNTR)
|
|
mdev->device_cap_flags |= IB_DEVICE_BAD_PKEY_CNTR;
|
|
|
|
if (dev_lim->flags & DEV_LIM_FLAG_BAD_QKEY_CNTR)
|
|
mdev->device_cap_flags |= IB_DEVICE_BAD_QKEY_CNTR;
|
|
|
|
if (dev_lim->flags & DEV_LIM_FLAG_RAW_MULTI)
|
|
mdev->device_cap_flags |= IB_DEVICE_RAW_MULTI;
|
|
|
|
if (dev_lim->flags & DEV_LIM_FLAG_AUTO_PATH_MIG)
|
|
mdev->device_cap_flags |= IB_DEVICE_AUTO_PATH_MIG;
|
|
|
|
if (dev_lim->flags & DEV_LIM_FLAG_UD_AV_PORT_ENFORCE)
|
|
mdev->device_cap_flags |= IB_DEVICE_UD_AV_PORT_ENFORCE;
|
|
|
|
if (dev_lim->flags & DEV_LIM_FLAG_SRQ)
|
|
mdev->mthca_flags |= MTHCA_FLAG_SRQ;
|
|
|
|
if (mthca_is_memfree(mdev))
|
|
if (dev_lim->flags & DEV_LIM_FLAG_IPOIB_CSUM)
|
|
mdev->device_cap_flags |= IB_DEVICE_UD_IP_CSUM;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int mthca_init_tavor(struct mthca_dev *mdev)
|
|
{
|
|
s64 size;
|
|
u8 status;
|
|
int err;
|
|
struct mthca_dev_lim dev_lim;
|
|
struct mthca_profile profile;
|
|
struct mthca_init_hca_param init_hca;
|
|
|
|
err = mthca_SYS_EN(mdev, &status);
|
|
if (err) {
|
|
mthca_err(mdev, "SYS_EN command failed, aborting.\n");
|
|
return err;
|
|
}
|
|
if (status) {
|
|
mthca_err(mdev, "SYS_EN returned status 0x%02x, "
|
|
"aborting.\n", status);
|
|
return -EINVAL;
|
|
}
|
|
|
|
err = mthca_QUERY_FW(mdev, &status);
|
|
if (err) {
|
|
mthca_err(mdev, "QUERY_FW command failed, aborting.\n");
|
|
goto err_disable;
|
|
}
|
|
if (status) {
|
|
mthca_err(mdev, "QUERY_FW returned status 0x%02x, "
|
|
"aborting.\n", status);
|
|
err = -EINVAL;
|
|
goto err_disable;
|
|
}
|
|
err = mthca_QUERY_DDR(mdev, &status);
|
|
if (err) {
|
|
mthca_err(mdev, "QUERY_DDR command failed, aborting.\n");
|
|
goto err_disable;
|
|
}
|
|
if (status) {
|
|
mthca_err(mdev, "QUERY_DDR returned status 0x%02x, "
|
|
"aborting.\n", status);
|
|
err = -EINVAL;
|
|
goto err_disable;
|
|
}
|
|
|
|
err = mthca_dev_lim(mdev, &dev_lim);
|
|
if (err) {
|
|
mthca_err(mdev, "QUERY_DEV_LIM command failed, aborting.\n");
|
|
goto err_disable;
|
|
}
|
|
|
|
profile = hca_profile;
|
|
profile.num_uar = dev_lim.uar_size / PAGE_SIZE;
|
|
profile.uarc_size = 0;
|
|
if (mdev->mthca_flags & MTHCA_FLAG_SRQ)
|
|
profile.num_srq = dev_lim.max_srqs;
|
|
|
|
size = mthca_make_profile(mdev, &profile, &dev_lim, &init_hca);
|
|
if (size < 0) {
|
|
err = size;
|
|
goto err_disable;
|
|
}
|
|
|
|
err = mthca_INIT_HCA(mdev, &init_hca, &status);
|
|
if (err) {
|
|
mthca_err(mdev, "INIT_HCA command failed, aborting.\n");
|
|
goto err_disable;
|
|
}
|
|
if (status) {
|
|
mthca_err(mdev, "INIT_HCA returned status 0x%02x, "
|
|
"aborting.\n", status);
|
|
err = -EINVAL;
|
|
goto err_disable;
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_disable:
|
|
mthca_SYS_DIS(mdev, &status);
|
|
|
|
return err;
|
|
}
|
|
|
|
static int mthca_load_fw(struct mthca_dev *mdev)
|
|
{
|
|
u8 status;
|
|
int err;
|
|
|
|
/* FIXME: use HCA-attached memory for FW if present */
|
|
|
|
mdev->fw.arbel.fw_icm =
|
|
mthca_alloc_icm(mdev, mdev->fw.arbel.fw_pages,
|
|
GFP_HIGHUSER | __GFP_NOWARN, 0);
|
|
if (!mdev->fw.arbel.fw_icm) {
|
|
mthca_err(mdev, "Couldn't allocate FW area, aborting.\n");
|
|
return -ENOMEM;
|
|
}
|
|
|
|
err = mthca_MAP_FA(mdev, mdev->fw.arbel.fw_icm, &status);
|
|
if (err) {
|
|
mthca_err(mdev, "MAP_FA command failed, aborting.\n");
|
|
goto err_free;
|
|
}
|
|
if (status) {
|
|
mthca_err(mdev, "MAP_FA returned status 0x%02x, aborting.\n", status);
|
|
err = -EINVAL;
|
|
goto err_free;
|
|
}
|
|
err = mthca_RUN_FW(mdev, &status);
|
|
if (err) {
|
|
mthca_err(mdev, "RUN_FW command failed, aborting.\n");
|
|
goto err_unmap_fa;
|
|
}
|
|
if (status) {
|
|
mthca_err(mdev, "RUN_FW returned status 0x%02x, aborting.\n", status);
|
|
err = -EINVAL;
|
|
goto err_unmap_fa;
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_unmap_fa:
|
|
mthca_UNMAP_FA(mdev, &status);
|
|
|
|
err_free:
|
|
mthca_free_icm(mdev, mdev->fw.arbel.fw_icm, 0);
|
|
return err;
|
|
}
|
|
|
|
static int mthca_init_icm(struct mthca_dev *mdev,
|
|
struct mthca_dev_lim *dev_lim,
|
|
struct mthca_init_hca_param *init_hca,
|
|
u64 icm_size)
|
|
{
|
|
u64 aux_pages;
|
|
u8 status;
|
|
int err;
|
|
|
|
err = mthca_SET_ICM_SIZE(mdev, icm_size, &aux_pages, &status);
|
|
if (err) {
|
|
mthca_err(mdev, "SET_ICM_SIZE command failed, aborting.\n");
|
|
return err;
|
|
}
|
|
if (status) {
|
|
mthca_err(mdev, "SET_ICM_SIZE returned status 0x%02x, "
|
|
"aborting.\n", status);
|
|
return -EINVAL;
|
|
}
|
|
|
|
mthca_dbg(mdev, "%lld KB of HCA context requires %lld KB aux memory.\n",
|
|
(unsigned long long) icm_size >> 10,
|
|
(unsigned long long) aux_pages << 2);
|
|
|
|
mdev->fw.arbel.aux_icm = mthca_alloc_icm(mdev, aux_pages,
|
|
GFP_HIGHUSER | __GFP_NOWARN, 0);
|
|
if (!mdev->fw.arbel.aux_icm) {
|
|
mthca_err(mdev, "Couldn't allocate aux memory, aborting.\n");
|
|
return -ENOMEM;
|
|
}
|
|
|
|
err = mthca_MAP_ICM_AUX(mdev, mdev->fw.arbel.aux_icm, &status);
|
|
if (err) {
|
|
mthca_err(mdev, "MAP_ICM_AUX command failed, aborting.\n");
|
|
goto err_free_aux;
|
|
}
|
|
if (status) {
|
|
mthca_err(mdev, "MAP_ICM_AUX returned status 0x%02x, aborting.\n", status);
|
|
err = -EINVAL;
|
|
goto err_free_aux;
|
|
}
|
|
|
|
err = mthca_map_eq_icm(mdev, init_hca->eqc_base);
|
|
if (err) {
|
|
mthca_err(mdev, "Failed to map EQ context memory, aborting.\n");
|
|
goto err_unmap_aux;
|
|
}
|
|
|
|
/* CPU writes to non-reserved MTTs, while HCA might DMA to reserved mtts */
|
|
mdev->limits.reserved_mtts = ALIGN(mdev->limits.reserved_mtts * mdev->limits.mtt_seg_size,
|
|
dma_get_cache_alignment()) / mdev->limits.mtt_seg_size;
|
|
|
|
mdev->mr_table.mtt_table = mthca_alloc_icm_table(mdev, init_hca->mtt_base,
|
|
mdev->limits.mtt_seg_size,
|
|
mdev->limits.num_mtt_segs,
|
|
mdev->limits.reserved_mtts,
|
|
1, 0);
|
|
if (!mdev->mr_table.mtt_table) {
|
|
mthca_err(mdev, "Failed to map MTT context memory, aborting.\n");
|
|
err = -ENOMEM;
|
|
goto err_unmap_eq;
|
|
}
|
|
|
|
mdev->mr_table.mpt_table = mthca_alloc_icm_table(mdev, init_hca->mpt_base,
|
|
dev_lim->mpt_entry_sz,
|
|
mdev->limits.num_mpts,
|
|
mdev->limits.reserved_mrws,
|
|
1, 1);
|
|
if (!mdev->mr_table.mpt_table) {
|
|
mthca_err(mdev, "Failed to map MPT context memory, aborting.\n");
|
|
err = -ENOMEM;
|
|
goto err_unmap_mtt;
|
|
}
|
|
|
|
mdev->qp_table.qp_table = mthca_alloc_icm_table(mdev, init_hca->qpc_base,
|
|
dev_lim->qpc_entry_sz,
|
|
mdev->limits.num_qps,
|
|
mdev->limits.reserved_qps,
|
|
0, 0);
|
|
if (!mdev->qp_table.qp_table) {
|
|
mthca_err(mdev, "Failed to map QP context memory, aborting.\n");
|
|
err = -ENOMEM;
|
|
goto err_unmap_mpt;
|
|
}
|
|
|
|
mdev->qp_table.eqp_table = mthca_alloc_icm_table(mdev, init_hca->eqpc_base,
|
|
dev_lim->eqpc_entry_sz,
|
|
mdev->limits.num_qps,
|
|
mdev->limits.reserved_qps,
|
|
0, 0);
|
|
if (!mdev->qp_table.eqp_table) {
|
|
mthca_err(mdev, "Failed to map EQP context memory, aborting.\n");
|
|
err = -ENOMEM;
|
|
goto err_unmap_qp;
|
|
}
|
|
|
|
mdev->qp_table.rdb_table = mthca_alloc_icm_table(mdev, init_hca->rdb_base,
|
|
MTHCA_RDB_ENTRY_SIZE,
|
|
mdev->limits.num_qps <<
|
|
mdev->qp_table.rdb_shift, 0,
|
|
0, 0);
|
|
if (!mdev->qp_table.rdb_table) {
|
|
mthca_err(mdev, "Failed to map RDB context memory, aborting\n");
|
|
err = -ENOMEM;
|
|
goto err_unmap_eqp;
|
|
}
|
|
|
|
mdev->cq_table.table = mthca_alloc_icm_table(mdev, init_hca->cqc_base,
|
|
dev_lim->cqc_entry_sz,
|
|
mdev->limits.num_cqs,
|
|
mdev->limits.reserved_cqs,
|
|
0, 0);
|
|
if (!mdev->cq_table.table) {
|
|
mthca_err(mdev, "Failed to map CQ context memory, aborting.\n");
|
|
err = -ENOMEM;
|
|
goto err_unmap_rdb;
|
|
}
|
|
|
|
if (mdev->mthca_flags & MTHCA_FLAG_SRQ) {
|
|
mdev->srq_table.table =
|
|
mthca_alloc_icm_table(mdev, init_hca->srqc_base,
|
|
dev_lim->srq_entry_sz,
|
|
mdev->limits.num_srqs,
|
|
mdev->limits.reserved_srqs,
|
|
0, 0);
|
|
if (!mdev->srq_table.table) {
|
|
mthca_err(mdev, "Failed to map SRQ context memory, "
|
|
"aborting.\n");
|
|
err = -ENOMEM;
|
|
goto err_unmap_cq;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* It's not strictly required, but for simplicity just map the
|
|
* whole multicast group table now. The table isn't very big
|
|
* and it's a lot easier than trying to track ref counts.
|
|
*/
|
|
mdev->mcg_table.table = mthca_alloc_icm_table(mdev, init_hca->mc_base,
|
|
MTHCA_MGM_ENTRY_SIZE,
|
|
mdev->limits.num_mgms +
|
|
mdev->limits.num_amgms,
|
|
mdev->limits.num_mgms +
|
|
mdev->limits.num_amgms,
|
|
0, 0);
|
|
if (!mdev->mcg_table.table) {
|
|
mthca_err(mdev, "Failed to map MCG context memory, aborting.\n");
|
|
err = -ENOMEM;
|
|
goto err_unmap_srq;
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_unmap_srq:
|
|
if (mdev->mthca_flags & MTHCA_FLAG_SRQ)
|
|
mthca_free_icm_table(mdev, mdev->srq_table.table);
|
|
|
|
err_unmap_cq:
|
|
mthca_free_icm_table(mdev, mdev->cq_table.table);
|
|
|
|
err_unmap_rdb:
|
|
mthca_free_icm_table(mdev, mdev->qp_table.rdb_table);
|
|
|
|
err_unmap_eqp:
|
|
mthca_free_icm_table(mdev, mdev->qp_table.eqp_table);
|
|
|
|
err_unmap_qp:
|
|
mthca_free_icm_table(mdev, mdev->qp_table.qp_table);
|
|
|
|
err_unmap_mpt:
|
|
mthca_free_icm_table(mdev, mdev->mr_table.mpt_table);
|
|
|
|
err_unmap_mtt:
|
|
mthca_free_icm_table(mdev, mdev->mr_table.mtt_table);
|
|
|
|
err_unmap_eq:
|
|
mthca_unmap_eq_icm(mdev);
|
|
|
|
err_unmap_aux:
|
|
mthca_UNMAP_ICM_AUX(mdev, &status);
|
|
|
|
err_free_aux:
|
|
mthca_free_icm(mdev, mdev->fw.arbel.aux_icm, 0);
|
|
|
|
return err;
|
|
}
|
|
|
|
static void mthca_free_icms(struct mthca_dev *mdev)
|
|
{
|
|
u8 status;
|
|
|
|
mthca_free_icm_table(mdev, mdev->mcg_table.table);
|
|
if (mdev->mthca_flags & MTHCA_FLAG_SRQ)
|
|
mthca_free_icm_table(mdev, mdev->srq_table.table);
|
|
mthca_free_icm_table(mdev, mdev->cq_table.table);
|
|
mthca_free_icm_table(mdev, mdev->qp_table.rdb_table);
|
|
mthca_free_icm_table(mdev, mdev->qp_table.eqp_table);
|
|
mthca_free_icm_table(mdev, mdev->qp_table.qp_table);
|
|
mthca_free_icm_table(mdev, mdev->mr_table.mpt_table);
|
|
mthca_free_icm_table(mdev, mdev->mr_table.mtt_table);
|
|
mthca_unmap_eq_icm(mdev);
|
|
|
|
mthca_UNMAP_ICM_AUX(mdev, &status);
|
|
mthca_free_icm(mdev, mdev->fw.arbel.aux_icm, 0);
|
|
}
|
|
|
|
static int mthca_init_arbel(struct mthca_dev *mdev)
|
|
{
|
|
struct mthca_dev_lim dev_lim;
|
|
struct mthca_profile profile;
|
|
struct mthca_init_hca_param init_hca;
|
|
s64 icm_size;
|
|
u8 status;
|
|
int err;
|
|
|
|
err = mthca_QUERY_FW(mdev, &status);
|
|
if (err) {
|
|
mthca_err(mdev, "QUERY_FW command failed, aborting.\n");
|
|
return err;
|
|
}
|
|
if (status) {
|
|
mthca_err(mdev, "QUERY_FW returned status 0x%02x, "
|
|
"aborting.\n", status);
|
|
return -EINVAL;
|
|
}
|
|
|
|
err = mthca_ENABLE_LAM(mdev, &status);
|
|
if (err) {
|
|
mthca_err(mdev, "ENABLE_LAM command failed, aborting.\n");
|
|
return err;
|
|
}
|
|
if (status == MTHCA_CMD_STAT_LAM_NOT_PRE) {
|
|
mthca_dbg(mdev, "No HCA-attached memory (running in MemFree mode)\n");
|
|
mdev->mthca_flags |= MTHCA_FLAG_NO_LAM;
|
|
} else if (status) {
|
|
mthca_err(mdev, "ENABLE_LAM returned status 0x%02x, "
|
|
"aborting.\n", status);
|
|
return -EINVAL;
|
|
}
|
|
|
|
err = mthca_load_fw(mdev);
|
|
if (err) {
|
|
mthca_err(mdev, "Failed to start FW, aborting.\n");
|
|
goto err_disable;
|
|
}
|
|
|
|
err = mthca_dev_lim(mdev, &dev_lim);
|
|
if (err) {
|
|
mthca_err(mdev, "QUERY_DEV_LIM command failed, aborting.\n");
|
|
goto err_stop_fw;
|
|
}
|
|
|
|
profile = hca_profile;
|
|
profile.num_uar = dev_lim.uar_size / PAGE_SIZE;
|
|
profile.num_udav = 0;
|
|
if (mdev->mthca_flags & MTHCA_FLAG_SRQ)
|
|
profile.num_srq = dev_lim.max_srqs;
|
|
|
|
icm_size = mthca_make_profile(mdev, &profile, &dev_lim, &init_hca);
|
|
if (icm_size < 0) {
|
|
err = icm_size;
|
|
goto err_stop_fw;
|
|
}
|
|
|
|
err = mthca_init_icm(mdev, &dev_lim, &init_hca, icm_size);
|
|
if (err)
|
|
goto err_stop_fw;
|
|
|
|
err = mthca_INIT_HCA(mdev, &init_hca, &status);
|
|
if (err) {
|
|
mthca_err(mdev, "INIT_HCA command failed, aborting.\n");
|
|
goto err_free_icm;
|
|
}
|
|
if (status) {
|
|
mthca_err(mdev, "INIT_HCA returned status 0x%02x, "
|
|
"aborting.\n", status);
|
|
err = -EINVAL;
|
|
goto err_free_icm;
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_free_icm:
|
|
mthca_free_icms(mdev);
|
|
|
|
err_stop_fw:
|
|
mthca_UNMAP_FA(mdev, &status);
|
|
mthca_free_icm(mdev, mdev->fw.arbel.fw_icm, 0);
|
|
|
|
err_disable:
|
|
if (!(mdev->mthca_flags & MTHCA_FLAG_NO_LAM))
|
|
mthca_DISABLE_LAM(mdev, &status);
|
|
|
|
return err;
|
|
}
|
|
|
|
static void mthca_close_hca(struct mthca_dev *mdev)
|
|
{
|
|
u8 status;
|
|
|
|
mthca_CLOSE_HCA(mdev, 0, &status);
|
|
|
|
if (mthca_is_memfree(mdev)) {
|
|
mthca_free_icms(mdev);
|
|
|
|
mthca_UNMAP_FA(mdev, &status);
|
|
mthca_free_icm(mdev, mdev->fw.arbel.fw_icm, 0);
|
|
|
|
if (!(mdev->mthca_flags & MTHCA_FLAG_NO_LAM))
|
|
mthca_DISABLE_LAM(mdev, &status);
|
|
} else
|
|
mthca_SYS_DIS(mdev, &status);
|
|
}
|
|
|
|
static int mthca_init_hca(struct mthca_dev *mdev)
|
|
{
|
|
u8 status;
|
|
int err;
|
|
struct mthca_adapter adapter;
|
|
|
|
if (mthca_is_memfree(mdev))
|
|
err = mthca_init_arbel(mdev);
|
|
else
|
|
err = mthca_init_tavor(mdev);
|
|
|
|
if (err)
|
|
return err;
|
|
|
|
err = mthca_QUERY_ADAPTER(mdev, &adapter, &status);
|
|
if (err) {
|
|
mthca_err(mdev, "QUERY_ADAPTER command failed, aborting.\n");
|
|
goto err_close;
|
|
}
|
|
if (status) {
|
|
mthca_err(mdev, "QUERY_ADAPTER returned status 0x%02x, "
|
|
"aborting.\n", status);
|
|
err = -EINVAL;
|
|
goto err_close;
|
|
}
|
|
|
|
mdev->eq_table.inta_pin = adapter.inta_pin;
|
|
if (!mthca_is_memfree(mdev))
|
|
mdev->rev_id = adapter.revision_id;
|
|
memcpy(mdev->board_id, adapter.board_id, sizeof mdev->board_id);
|
|
|
|
return 0;
|
|
|
|
err_close:
|
|
mthca_close_hca(mdev);
|
|
return err;
|
|
}
|
|
|
|
static int mthca_setup_hca(struct mthca_dev *dev)
|
|
{
|
|
int err;
|
|
u8 status;
|
|
|
|
MTHCA_INIT_DOORBELL_LOCK(&dev->doorbell_lock);
|
|
|
|
err = mthca_init_uar_table(dev);
|
|
if (err) {
|
|
mthca_err(dev, "Failed to initialize "
|
|
"user access region table, aborting.\n");
|
|
return err;
|
|
}
|
|
|
|
err = mthca_uar_alloc(dev, &dev->driver_uar);
|
|
if (err) {
|
|
mthca_err(dev, "Failed to allocate driver access region, "
|
|
"aborting.\n");
|
|
goto err_uar_table_free;
|
|
}
|
|
|
|
dev->kar = ioremap(dev->driver_uar.pfn << PAGE_SHIFT, PAGE_SIZE);
|
|
if (!dev->kar) {
|
|
mthca_err(dev, "Couldn't map kernel access region, "
|
|
"aborting.\n");
|
|
err = -ENOMEM;
|
|
goto err_uar_free;
|
|
}
|
|
|
|
err = mthca_init_pd_table(dev);
|
|
if (err) {
|
|
mthca_err(dev, "Failed to initialize "
|
|
"protection domain table, aborting.\n");
|
|
goto err_kar_unmap;
|
|
}
|
|
|
|
err = mthca_init_mr_table(dev);
|
|
if (err) {
|
|
mthca_err(dev, "Failed to initialize "
|
|
"memory region table, aborting.\n");
|
|
goto err_pd_table_free;
|
|
}
|
|
|
|
err = mthca_pd_alloc(dev, 1, &dev->driver_pd);
|
|
if (err) {
|
|
mthca_err(dev, "Failed to create driver PD, "
|
|
"aborting.\n");
|
|
goto err_mr_table_free;
|
|
}
|
|
|
|
err = mthca_init_eq_table(dev);
|
|
if (err) {
|
|
mthca_err(dev, "Failed to initialize "
|
|
"event queue table, aborting.\n");
|
|
goto err_pd_free;
|
|
}
|
|
|
|
err = mthca_cmd_use_events(dev);
|
|
if (err) {
|
|
mthca_err(dev, "Failed to switch to event-driven "
|
|
"firmware commands, aborting.\n");
|
|
goto err_eq_table_free;
|
|
}
|
|
|
|
err = mthca_NOP(dev, &status);
|
|
if (err || status) {
|
|
if (dev->mthca_flags & MTHCA_FLAG_MSI_X) {
|
|
mthca_warn(dev, "NOP command failed to generate interrupt "
|
|
"(IRQ %d).\n",
|
|
dev->eq_table.eq[MTHCA_EQ_CMD].msi_x_vector);
|
|
mthca_warn(dev, "Trying again with MSI-X disabled.\n");
|
|
} else {
|
|
mthca_err(dev, "NOP command failed to generate interrupt "
|
|
"(IRQ %d), aborting.\n",
|
|
dev->pdev->irq);
|
|
mthca_err(dev, "BIOS or ACPI interrupt routing problem?\n");
|
|
}
|
|
|
|
goto err_cmd_poll;
|
|
}
|
|
|
|
mthca_dbg(dev, "NOP command IRQ test passed\n");
|
|
|
|
err = mthca_init_cq_table(dev);
|
|
if (err) {
|
|
mthca_err(dev, "Failed to initialize "
|
|
"completion queue table, aborting.\n");
|
|
goto err_cmd_poll;
|
|
}
|
|
|
|
err = mthca_init_srq_table(dev);
|
|
if (err) {
|
|
mthca_err(dev, "Failed to initialize "
|
|
"shared receive queue table, aborting.\n");
|
|
goto err_cq_table_free;
|
|
}
|
|
|
|
err = mthca_init_qp_table(dev);
|
|
if (err) {
|
|
mthca_err(dev, "Failed to initialize "
|
|
"queue pair table, aborting.\n");
|
|
goto err_srq_table_free;
|
|
}
|
|
|
|
err = mthca_init_av_table(dev);
|
|
if (err) {
|
|
mthca_err(dev, "Failed to initialize "
|
|
"address vector table, aborting.\n");
|
|
goto err_qp_table_free;
|
|
}
|
|
|
|
err = mthca_init_mcg_table(dev);
|
|
if (err) {
|
|
mthca_err(dev, "Failed to initialize "
|
|
"multicast group table, aborting.\n");
|
|
goto err_av_table_free;
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_av_table_free:
|
|
mthca_cleanup_av_table(dev);
|
|
|
|
err_qp_table_free:
|
|
mthca_cleanup_qp_table(dev);
|
|
|
|
err_srq_table_free:
|
|
mthca_cleanup_srq_table(dev);
|
|
|
|
err_cq_table_free:
|
|
mthca_cleanup_cq_table(dev);
|
|
|
|
err_cmd_poll:
|
|
mthca_cmd_use_polling(dev);
|
|
|
|
err_eq_table_free:
|
|
mthca_cleanup_eq_table(dev);
|
|
|
|
err_pd_free:
|
|
mthca_pd_free(dev, &dev->driver_pd);
|
|
|
|
err_mr_table_free:
|
|
mthca_cleanup_mr_table(dev);
|
|
|
|
err_pd_table_free:
|
|
mthca_cleanup_pd_table(dev);
|
|
|
|
err_kar_unmap:
|
|
iounmap(dev->kar);
|
|
|
|
err_uar_free:
|
|
mthca_uar_free(dev, &dev->driver_uar);
|
|
|
|
err_uar_table_free:
|
|
mthca_cleanup_uar_table(dev);
|
|
return err;
|
|
}
|
|
|
|
static int mthca_enable_msi_x(struct mthca_dev *mdev)
|
|
{
|
|
struct msix_entry entries[3];
|
|
int err;
|
|
|
|
entries[0].entry = 0;
|
|
entries[1].entry = 1;
|
|
entries[2].entry = 2;
|
|
|
|
err = pci_enable_msix(mdev->pdev, entries, ARRAY_SIZE(entries));
|
|
if (err) {
|
|
if (err > 0)
|
|
mthca_info(mdev, "Only %d MSI-X vectors available, "
|
|
"not using MSI-X\n", err);
|
|
return err;
|
|
}
|
|
|
|
mdev->eq_table.eq[MTHCA_EQ_COMP ].msi_x_vector = entries[0].vector;
|
|
mdev->eq_table.eq[MTHCA_EQ_ASYNC].msi_x_vector = entries[1].vector;
|
|
mdev->eq_table.eq[MTHCA_EQ_CMD ].msi_x_vector = entries[2].vector;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Types of supported HCA */
|
|
enum {
|
|
TAVOR, /* MT23108 */
|
|
ARBEL_COMPAT, /* MT25208 in Tavor compat mode */
|
|
ARBEL_NATIVE, /* MT25208 with extended features */
|
|
SINAI /* MT25204 */
|
|
};
|
|
|
|
#define MTHCA_FW_VER(major, minor, subminor) \
|
|
(((u64) (major) << 32) | ((u64) (minor) << 16) | (u64) (subminor))
|
|
|
|
static struct {
|
|
u64 latest_fw;
|
|
u32 flags;
|
|
} mthca_hca_table[] = {
|
|
[TAVOR] = { .latest_fw = MTHCA_FW_VER(3, 5, 0),
|
|
.flags = 0 },
|
|
[ARBEL_COMPAT] = { .latest_fw = MTHCA_FW_VER(4, 8, 200),
|
|
.flags = MTHCA_FLAG_PCIE },
|
|
[ARBEL_NATIVE] = { .latest_fw = MTHCA_FW_VER(5, 3, 0),
|
|
.flags = MTHCA_FLAG_MEMFREE |
|
|
MTHCA_FLAG_PCIE },
|
|
[SINAI] = { .latest_fw = MTHCA_FW_VER(1, 2, 0),
|
|
.flags = MTHCA_FLAG_MEMFREE |
|
|
MTHCA_FLAG_PCIE |
|
|
MTHCA_FLAG_SINAI_OPT }
|
|
};
|
|
|
|
static int __mthca_init_one(struct pci_dev *pdev, int hca_type)
|
|
{
|
|
int ddr_hidden = 0;
|
|
int err;
|
|
struct mthca_dev *mdev;
|
|
|
|
printk(KERN_INFO PFX "Initializing %s\n",
|
|
pci_name(pdev));
|
|
|
|
err = pci_enable_device(pdev);
|
|
if (err) {
|
|
dev_err(&pdev->dev, "Cannot enable PCI device, "
|
|
"aborting.\n");
|
|
return err;
|
|
}
|
|
|
|
/*
|
|
* Check for BARs. We expect 0: 1MB, 2: 8MB, 4: DDR (may not
|
|
* be present)
|
|
*/
|
|
if (!(pci_resource_flags(pdev, 0) & IORESOURCE_MEM) ||
|
|
pci_resource_len(pdev, 0) != 1 << 20) {
|
|
dev_err(&pdev->dev, "Missing DCS, aborting.\n");
|
|
err = -ENODEV;
|
|
goto err_disable_pdev;
|
|
}
|
|
if (!(pci_resource_flags(pdev, 2) & IORESOURCE_MEM)) {
|
|
dev_err(&pdev->dev, "Missing UAR, aborting.\n");
|
|
err = -ENODEV;
|
|
goto err_disable_pdev;
|
|
}
|
|
if (!(pci_resource_flags(pdev, 4) & IORESOURCE_MEM))
|
|
ddr_hidden = 1;
|
|
|
|
err = pci_request_regions(pdev, DRV_NAME);
|
|
if (err) {
|
|
dev_err(&pdev->dev, "Cannot obtain PCI resources, "
|
|
"aborting.\n");
|
|
goto err_disable_pdev;
|
|
}
|
|
|
|
pci_set_master(pdev);
|
|
|
|
err = pci_set_dma_mask(pdev, DMA_BIT_MASK(64));
|
|
if (err) {
|
|
dev_warn(&pdev->dev, "Warning: couldn't set 64-bit PCI DMA mask.\n");
|
|
err = pci_set_dma_mask(pdev, DMA_BIT_MASK(32));
|
|
if (err) {
|
|
dev_err(&pdev->dev, "Can't set PCI DMA mask, aborting.\n");
|
|
goto err_free_res;
|
|
}
|
|
}
|
|
err = pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(64));
|
|
if (err) {
|
|
dev_warn(&pdev->dev, "Warning: couldn't set 64-bit "
|
|
"consistent PCI DMA mask.\n");
|
|
err = pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(32));
|
|
if (err) {
|
|
dev_err(&pdev->dev, "Can't set consistent PCI DMA mask, "
|
|
"aborting.\n");
|
|
goto err_free_res;
|
|
}
|
|
}
|
|
|
|
mdev = (struct mthca_dev *) ib_alloc_device(sizeof *mdev);
|
|
if (!mdev) {
|
|
dev_err(&pdev->dev, "Device struct alloc failed, "
|
|
"aborting.\n");
|
|
err = -ENOMEM;
|
|
goto err_free_res;
|
|
}
|
|
|
|
mdev->pdev = pdev;
|
|
|
|
mdev->mthca_flags = mthca_hca_table[hca_type].flags;
|
|
if (ddr_hidden)
|
|
mdev->mthca_flags |= MTHCA_FLAG_DDR_HIDDEN;
|
|
|
|
/*
|
|
* Now reset the HCA before we touch the PCI capabilities or
|
|
* attempt a firmware command, since a boot ROM may have left
|
|
* the HCA in an undefined state.
|
|
*/
|
|
err = mthca_reset(mdev);
|
|
if (err) {
|
|
mthca_err(mdev, "Failed to reset HCA, aborting.\n");
|
|
goto err_free_dev;
|
|
}
|
|
|
|
if (mthca_cmd_init(mdev)) {
|
|
mthca_err(mdev, "Failed to init command interface, aborting.\n");
|
|
goto err_free_dev;
|
|
}
|
|
|
|
err = mthca_tune_pci(mdev);
|
|
if (err)
|
|
goto err_cmd;
|
|
|
|
err = mthca_init_hca(mdev);
|
|
if (err)
|
|
goto err_cmd;
|
|
|
|
if (mdev->fw_ver < mthca_hca_table[hca_type].latest_fw) {
|
|
mthca_warn(mdev, "HCA FW version %d.%d.%03d is old (%d.%d.%03d is current).\n",
|
|
(int) (mdev->fw_ver >> 32), (int) (mdev->fw_ver >> 16) & 0xffff,
|
|
(int) (mdev->fw_ver & 0xffff),
|
|
(int) (mthca_hca_table[hca_type].latest_fw >> 32),
|
|
(int) (mthca_hca_table[hca_type].latest_fw >> 16) & 0xffff,
|
|
(int) (mthca_hca_table[hca_type].latest_fw & 0xffff));
|
|
mthca_warn(mdev, "If you have problems, try updating your HCA FW.\n");
|
|
}
|
|
|
|
if (msi_x && !mthca_enable_msi_x(mdev))
|
|
mdev->mthca_flags |= MTHCA_FLAG_MSI_X;
|
|
|
|
err = mthca_setup_hca(mdev);
|
|
if (err == -EBUSY && (mdev->mthca_flags & MTHCA_FLAG_MSI_X)) {
|
|
if (mdev->mthca_flags & MTHCA_FLAG_MSI_X)
|
|
pci_disable_msix(pdev);
|
|
mdev->mthca_flags &= ~MTHCA_FLAG_MSI_X;
|
|
|
|
err = mthca_setup_hca(mdev);
|
|
}
|
|
|
|
if (err)
|
|
goto err_close;
|
|
|
|
err = mthca_register_device(mdev);
|
|
if (err)
|
|
goto err_cleanup;
|
|
|
|
err = mthca_create_agents(mdev);
|
|
if (err)
|
|
goto err_unregister;
|
|
|
|
pci_set_drvdata(pdev, mdev);
|
|
mdev->hca_type = hca_type;
|
|
|
|
mdev->active = true;
|
|
|
|
return 0;
|
|
|
|
err_unregister:
|
|
mthca_unregister_device(mdev);
|
|
|
|
err_cleanup:
|
|
mthca_cleanup_mcg_table(mdev);
|
|
mthca_cleanup_av_table(mdev);
|
|
mthca_cleanup_qp_table(mdev);
|
|
mthca_cleanup_srq_table(mdev);
|
|
mthca_cleanup_cq_table(mdev);
|
|
mthca_cmd_use_polling(mdev);
|
|
mthca_cleanup_eq_table(mdev);
|
|
|
|
mthca_pd_free(mdev, &mdev->driver_pd);
|
|
|
|
mthca_cleanup_mr_table(mdev);
|
|
mthca_cleanup_pd_table(mdev);
|
|
mthca_cleanup_uar_table(mdev);
|
|
|
|
err_close:
|
|
if (mdev->mthca_flags & MTHCA_FLAG_MSI_X)
|
|
pci_disable_msix(pdev);
|
|
|
|
mthca_close_hca(mdev);
|
|
|
|
err_cmd:
|
|
mthca_cmd_cleanup(mdev);
|
|
|
|
err_free_dev:
|
|
ib_dealloc_device(&mdev->ib_dev);
|
|
|
|
err_free_res:
|
|
pci_release_regions(pdev);
|
|
|
|
err_disable_pdev:
|
|
pci_disable_device(pdev);
|
|
pci_set_drvdata(pdev, NULL);
|
|
return err;
|
|
}
|
|
|
|
static void __mthca_remove_one(struct pci_dev *pdev)
|
|
{
|
|
struct mthca_dev *mdev = pci_get_drvdata(pdev);
|
|
u8 status;
|
|
int p;
|
|
|
|
if (mdev) {
|
|
mthca_free_agents(mdev);
|
|
mthca_unregister_device(mdev);
|
|
|
|
for (p = 1; p <= mdev->limits.num_ports; ++p)
|
|
mthca_CLOSE_IB(mdev, p, &status);
|
|
|
|
mthca_cleanup_mcg_table(mdev);
|
|
mthca_cleanup_av_table(mdev);
|
|
mthca_cleanup_qp_table(mdev);
|
|
mthca_cleanup_srq_table(mdev);
|
|
mthca_cleanup_cq_table(mdev);
|
|
mthca_cmd_use_polling(mdev);
|
|
mthca_cleanup_eq_table(mdev);
|
|
|
|
mthca_pd_free(mdev, &mdev->driver_pd);
|
|
|
|
mthca_cleanup_mr_table(mdev);
|
|
mthca_cleanup_pd_table(mdev);
|
|
|
|
iounmap(mdev->kar);
|
|
mthca_uar_free(mdev, &mdev->driver_uar);
|
|
mthca_cleanup_uar_table(mdev);
|
|
mthca_close_hca(mdev);
|
|
mthca_cmd_cleanup(mdev);
|
|
|
|
if (mdev->mthca_flags & MTHCA_FLAG_MSI_X)
|
|
pci_disable_msix(pdev);
|
|
|
|
ib_dealloc_device(&mdev->ib_dev);
|
|
pci_release_regions(pdev);
|
|
pci_disable_device(pdev);
|
|
pci_set_drvdata(pdev, NULL);
|
|
}
|
|
}
|
|
|
|
int __mthca_restart_one(struct pci_dev *pdev)
|
|
{
|
|
struct mthca_dev *mdev;
|
|
int hca_type;
|
|
|
|
mdev = pci_get_drvdata(pdev);
|
|
if (!mdev)
|
|
return -ENODEV;
|
|
hca_type = mdev->hca_type;
|
|
__mthca_remove_one(pdev);
|
|
return __mthca_init_one(pdev, hca_type);
|
|
}
|
|
|
|
static int __devinit mthca_init_one(struct pci_dev *pdev,
|
|
const struct pci_device_id *id)
|
|
{
|
|
int ret;
|
|
|
|
mutex_lock(&mthca_device_mutex);
|
|
|
|
printk_once(KERN_INFO "%s", mthca_version);
|
|
|
|
if (id->driver_data >= ARRAY_SIZE(mthca_hca_table)) {
|
|
printk(KERN_ERR PFX "%s has invalid driver data %lx\n",
|
|
pci_name(pdev), id->driver_data);
|
|
mutex_unlock(&mthca_device_mutex);
|
|
return -ENODEV;
|
|
}
|
|
|
|
ret = __mthca_init_one(pdev, id->driver_data);
|
|
|
|
mutex_unlock(&mthca_device_mutex);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void __devexit mthca_remove_one(struct pci_dev *pdev)
|
|
{
|
|
mutex_lock(&mthca_device_mutex);
|
|
__mthca_remove_one(pdev);
|
|
mutex_unlock(&mthca_device_mutex);
|
|
}
|
|
|
|
static struct pci_device_id mthca_pci_table[] = {
|
|
{ PCI_DEVICE(PCI_VENDOR_ID_MELLANOX, PCI_DEVICE_ID_MELLANOX_TAVOR),
|
|
.driver_data = TAVOR },
|
|
{ PCI_DEVICE(PCI_VENDOR_ID_TOPSPIN, PCI_DEVICE_ID_MELLANOX_TAVOR),
|
|
.driver_data = TAVOR },
|
|
{ PCI_DEVICE(PCI_VENDOR_ID_MELLANOX, PCI_DEVICE_ID_MELLANOX_ARBEL_COMPAT),
|
|
.driver_data = ARBEL_COMPAT },
|
|
{ PCI_DEVICE(PCI_VENDOR_ID_TOPSPIN, PCI_DEVICE_ID_MELLANOX_ARBEL_COMPAT),
|
|
.driver_data = ARBEL_COMPAT },
|
|
{ PCI_DEVICE(PCI_VENDOR_ID_MELLANOX, PCI_DEVICE_ID_MELLANOX_ARBEL),
|
|
.driver_data = ARBEL_NATIVE },
|
|
{ PCI_DEVICE(PCI_VENDOR_ID_TOPSPIN, PCI_DEVICE_ID_MELLANOX_ARBEL),
|
|
.driver_data = ARBEL_NATIVE },
|
|
{ PCI_DEVICE(PCI_VENDOR_ID_MELLANOX, PCI_DEVICE_ID_MELLANOX_SINAI),
|
|
.driver_data = SINAI },
|
|
{ PCI_DEVICE(PCI_VENDOR_ID_TOPSPIN, PCI_DEVICE_ID_MELLANOX_SINAI),
|
|
.driver_data = SINAI },
|
|
{ PCI_DEVICE(PCI_VENDOR_ID_MELLANOX, PCI_DEVICE_ID_MELLANOX_SINAI_OLD),
|
|
.driver_data = SINAI },
|
|
{ PCI_DEVICE(PCI_VENDOR_ID_TOPSPIN, PCI_DEVICE_ID_MELLANOX_SINAI_OLD),
|
|
.driver_data = SINAI },
|
|
{ 0, }
|
|
};
|
|
|
|
MODULE_DEVICE_TABLE(pci, mthca_pci_table);
|
|
|
|
static struct pci_driver mthca_driver = {
|
|
.name = DRV_NAME,
|
|
.id_table = mthca_pci_table,
|
|
.probe = mthca_init_one,
|
|
.remove = __devexit_p(mthca_remove_one)
|
|
};
|
|
|
|
static void __init __mthca_check_profile_val(const char *name, int *pval,
|
|
int pval_default)
|
|
{
|
|
/* value must be positive and power of 2 */
|
|
int old_pval = *pval;
|
|
|
|
if (old_pval <= 0)
|
|
*pval = pval_default;
|
|
else
|
|
*pval = roundup_pow_of_two(old_pval);
|
|
|
|
if (old_pval != *pval) {
|
|
printk(KERN_WARNING PFX "Invalid value %d for %s in module parameter.\n",
|
|
old_pval, name);
|
|
printk(KERN_WARNING PFX "Corrected %s to %d.\n", name, *pval);
|
|
}
|
|
}
|
|
|
|
#define mthca_check_profile_val(name, default) \
|
|
__mthca_check_profile_val(#name, &hca_profile.name, default)
|
|
|
|
static void __init mthca_validate_profile(void)
|
|
{
|
|
mthca_check_profile_val(num_qp, MTHCA_DEFAULT_NUM_QP);
|
|
mthca_check_profile_val(rdb_per_qp, MTHCA_DEFAULT_RDB_PER_QP);
|
|
mthca_check_profile_val(num_cq, MTHCA_DEFAULT_NUM_CQ);
|
|
mthca_check_profile_val(num_mcg, MTHCA_DEFAULT_NUM_MCG);
|
|
mthca_check_profile_val(num_mpt, MTHCA_DEFAULT_NUM_MPT);
|
|
mthca_check_profile_val(num_mtt, MTHCA_DEFAULT_NUM_MTT);
|
|
mthca_check_profile_val(num_udav, MTHCA_DEFAULT_NUM_UDAV);
|
|
mthca_check_profile_val(fmr_reserved_mtts, MTHCA_DEFAULT_NUM_RESERVED_MTTS);
|
|
|
|
if (hca_profile.fmr_reserved_mtts >= hca_profile.num_mtt) {
|
|
printk(KERN_WARNING PFX "Invalid fmr_reserved_mtts module parameter %d.\n",
|
|
hca_profile.fmr_reserved_mtts);
|
|
printk(KERN_WARNING PFX "(Must be smaller than num_mtt %d)\n",
|
|
hca_profile.num_mtt);
|
|
hca_profile.fmr_reserved_mtts = hca_profile.num_mtt / 2;
|
|
printk(KERN_WARNING PFX "Corrected fmr_reserved_mtts to %d.\n",
|
|
hca_profile.fmr_reserved_mtts);
|
|
}
|
|
|
|
if ((log_mtts_per_seg < 1) || (log_mtts_per_seg > 5)) {
|
|
printk(KERN_WARNING PFX "bad log_mtts_per_seg (%d). Using default - %d\n",
|
|
log_mtts_per_seg, ilog2(MTHCA_MTT_SEG_SIZE / 8));
|
|
log_mtts_per_seg = ilog2(MTHCA_MTT_SEG_SIZE / 8);
|
|
}
|
|
}
|
|
|
|
static int __init mthca_init(void)
|
|
{
|
|
int ret;
|
|
|
|
mthca_validate_profile();
|
|
|
|
ret = mthca_catas_init();
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = pci_register_driver(&mthca_driver);
|
|
if (ret < 0) {
|
|
mthca_catas_cleanup();
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void __exit mthca_cleanup(void)
|
|
{
|
|
pci_unregister_driver(&mthca_driver);
|
|
mthca_catas_cleanup();
|
|
}
|
|
|
|
module_init(mthca_init);
|
|
module_exit(mthca_cleanup);
|