summaryrefslogtreecommitdiffstats
path: root/contrib/syslinux-4.02/gpxe/src/drivers/net/mtnic.c
diff options
context:
space:
mode:
Diffstat (limited to 'contrib/syslinux-4.02/gpxe/src/drivers/net/mtnic.c')
-rw-r--r--contrib/syslinux-4.02/gpxe/src/drivers/net/mtnic.c1853
1 files changed, 1853 insertions, 0 deletions
diff --git a/contrib/syslinux-4.02/gpxe/src/drivers/net/mtnic.c b/contrib/syslinux-4.02/gpxe/src/drivers/net/mtnic.c
new file mode 100644
index 0000000..d7ee8d2
--- /dev/null
+++ b/contrib/syslinux-4.02/gpxe/src/drivers/net/mtnic.c
@@ -0,0 +1,1853 @@
+/*
+ * Copyright (c) 2007 Mellanox Technologies. All rights reserved.
+ *
+ * This software is available to you under a choice of one of two
+ * licenses. You may choose to be licensed under the terms of the GNU
+ * General Public License (GPL) Version 2, available from the file
+ * COPYING in the main directory of this source tree, or the
+ * OpenIB.org BSD license below:
+ *
+ * Redistribution and use in source and binary forms, with or
+ * without modification, are permitted provided that the following
+ * conditions are met:
+ *
+ * - Redistributions of source code must retain the above
+ * copyright notice, this list of conditions and the following
+ * disclaimer.
+ *
+ * - Redistributions in binary form must reproduce the above
+ * copyright notice, this list of conditions and the following
+ * disclaimer in the documentation and/or other materials
+ * provided with the distribution.
+ *
+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
+ * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
+ * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
+ * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
+ * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
+ * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
+ * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
+ * SOFTWARE.
+ *
+ */
+
+FILE_LICENCE ( GPL2_ONLY );
+
+#include <strings.h>
+#include <errno.h>
+#include <gpxe/malloc.h>
+#include <gpxe/umalloc.h>
+#include <byteswap.h>
+#include <unistd.h>
+#include <gpxe/io.h>
+#include <gpxe/pci.h>
+#include <gpxe/ethernet.h>
+#include <gpxe/netdevice.h>
+#include <gpxe/iobuf.h>
+#include "mtnic.h"
+
+
+/*
+
+
+ mtnic.c - gPXE driver for Mellanox 10Gig ConnectX EN
+
+
+*/
+
+
+
+/********************************************************************
+*
+* MTNIC allocation functions
+*
+*********************************************************************/
+/**
+* mtnic_alloc_aligned
+*
+* @v unsigned int size size
+* @v void **va virtual address
+* @v u32 *pa physical address
+* @v u32 aligment aligment
+*
+* Function allocate aligned buffer and put it's virtual address in 'va'
+* and it's physical aligned address in 'pa'
+*/
+static int
+mtnic_alloc_aligned(unsigned int size, void **va, unsigned long *pa, unsigned int alignment)
+{
+ *va = alloc_memblock(size, alignment);
+ if (!*va) {
+ return -EADDRINUSE;
+ }
+ *pa = (u32)virt_to_bus(*va);
+ return 0;
+}
+
+
+
+/**
+ *
+ * mtnic alloc command interface
+ *
+ */
+static int
+mtnic_alloc_cmdif(struct mtnic *mtnic)
+{
+ u32 bar = mtnic_pci_dev.dev.bar[0];
+
+ mtnic->hcr = ioremap(bar + MTNIC_HCR_BASE, MTNIC_HCR_SIZE);
+ if ( !mtnic->hcr ) {
+ DBG("Couldn't map command register\n");
+ return -EADDRINUSE;
+ }
+ mtnic_alloc_aligned(PAGE_SIZE, (void *)&mtnic->cmd.buf, &mtnic->cmd.mapping, PAGE_SIZE);
+ if ( !mtnic->cmd.buf ) {
+ DBG("Error in allocating buffer for command interface\n");
+ return -EADDRINUSE;
+ }
+ return 0;
+}
+
+/**
+ * Free RX io buffers
+ */
+static void
+mtnic_free_io_buffers(struct mtnic_ring *ring)
+{
+ int index;
+
+ for (; ring->cons <= ring->prod; ++ring->cons) {
+ index = ring->cons & ring->size_mask;
+ if ( ring->iobuf[index] ) {
+ free_iob(ring->iobuf[index]);
+ }
+ }
+}
+
+
+
+/**
+ *
+ * mtnic alloc and attach io buffers
+ *
+ */
+static int
+mtnic_alloc_iobuf(struct mtnic_port *priv, struct mtnic_ring *ring,
+ unsigned int size)
+{
+ struct mtnic_rx_desc *rx_desc_ptr = ring->buf;
+ u32 index;
+
+ while ((u32)(ring->prod - ring->cons) < UNITS_BUFFER_SIZE) {
+ index = ring->prod & ring->size_mask;
+ ring->iobuf[index] = alloc_iob(size);
+ if (!ring->iobuf[index]) {
+ if (ring->prod <= (ring->cons + 1)) {
+ DBG ( "Dropping packet, buffer is full\n" );
+ }
+ break;
+ }
+
+ /* Attach io_buffer to descriptor */
+ rx_desc_ptr = ring->buf +
+ (sizeof(struct mtnic_rx_desc) * index);
+ rx_desc_ptr->data.count = cpu_to_be32(size);
+ rx_desc_ptr->data.mem_type = priv->mtnic->fw.mem_type_snoop_be;
+ rx_desc_ptr->data.addr_l = cpu_to_be32(
+ virt_to_bus(ring->iobuf[index]->data));
+
+ ++ ring->prod;
+ }
+
+ /* Update RX producer index (PI) */
+ ring->db->count = cpu_to_be32(ring->prod & 0xffff);
+ return 0;
+}
+
+
+/**
+ * mtnic alloc ring
+ *
+ * Alloc and configure TX or RX ring
+ *
+ */
+static int
+mtnic_alloc_ring(struct mtnic_port *priv, struct mtnic_ring *ring,
+ u32 size, u16 stride, u16 cq, u8 is_rx)
+{
+ unsigned int i;
+ int err;
+ struct mtnic_rx_desc *rx_desc;
+ struct mtnic_tx_desc *tx_desc;
+
+ ring->size = size; /* Number of descriptors */
+ ring->size_mask = size - 1;
+ ring->stride = stride; /* Size of each entry */
+ ring->cq = cq; /* CQ number associated with this ring */
+ ring->cons = 0;
+ ring->prod = 0;
+
+ /* Alloc descriptors buffer */
+ ring->buf_size = ring->size * ((is_rx) ? sizeof(struct mtnic_rx_desc) :
+ sizeof(struct mtnic_tx_desc));
+ err = mtnic_alloc_aligned(ring->buf_size, (void *)&ring->buf,
+ &ring->dma, PAGE_SIZE);
+ if (err) {
+ DBG("Failed allocating descriptor ring sizeof %x\n",
+ ring->buf_size);
+ return -EADDRINUSE;
+ }
+ memset(ring->buf, 0, ring->buf_size);
+
+ DBG("Allocated %s ring (addr:%p) - buf:%p size:%x"
+ "buf_size:%x dma:%lx\n",
+ is_rx ? "Rx" : "Tx", ring, ring->buf, ring->size,
+ ring->buf_size, ring->dma);
+
+
+ if (is_rx) { /* RX ring */
+ /* Alloc doorbell */
+ err = mtnic_alloc_aligned(sizeof(struct mtnic_cq_db_record),
+ (void *)&ring->db, &ring->db_dma, 32);
+ if (err) {
+ DBG("Failed allocating Rx ring doorbell record\n");
+ free_memblock(ring->buf, ring->buf_size);
+ return -EADDRINUSE;
+ }
+
+ /* ==- Configure Descriptor -== */
+ /* Init ctrl seg of rx desc */
+ for (i = 0; i < UNITS_BUFFER_SIZE; ++i) {
+ rx_desc = ring->buf +
+ (sizeof(struct mtnic_rx_desc) * i);
+ /* Pre-link descriptor */
+ rx_desc->next = cpu_to_be16(i + 1);
+ }
+ /*The last ctrl descriptor is '0' and points to the first one*/
+
+ /* Alloc IO_BUFFERS */
+ err = mtnic_alloc_iobuf ( priv, ring, DEF_IOBUF_SIZE );
+ if (err) {
+ DBG("ERROR Allocating io buffer\n");
+ free_memblock(ring->buf, ring->buf_size);
+ return -EADDRINUSE;
+ }
+
+ } else { /* TX ring */
+ /* Set initial ownership of all Tx Desc' to SW (1) */
+ for (i = 0; i < ring->size; i++) {
+ tx_desc = ring->buf + ring->stride * i;
+ tx_desc->ctrl.op_own = cpu_to_be32(MTNIC_BIT_DESC_OWN);
+ }
+ /* DB */
+ ring->db_offset = cpu_to_be32(
+ ((u32) priv->mtnic->fw.tx_offset[priv->port]) << 8);
+
+ /* Map Tx+CQ doorbells */
+ DBG("Mapping TxCQ doorbell at offset:0x%x\n",
+ priv->mtnic->fw.txcq_db_offset);
+ ring->txcq_db = ioremap(mtnic_pci_dev.dev.bar[2] +
+ priv->mtnic->fw.txcq_db_offset, PAGE_SIZE);
+ if (!ring->txcq_db) {
+ DBG("Couldn't map txcq doorbell, aborting...\n");
+ free_memblock(ring->buf, ring->buf_size);
+ return -EADDRINUSE;
+ }
+ }
+
+ return 0;
+}
+
+
+
+/**
+ * mtnic alloc CQ
+ *
+ * Alloc and configure CQ.
+ *
+ */
+static int
+mtnic_alloc_cq(struct net_device *dev, int num, struct mtnic_cq *cq,
+ u8 is_rx, u32 size, u32 offset_ind)
+{
+ int err ;
+ unsigned int i;
+
+ cq->num = num;
+ cq->dev = dev;
+ cq->size = size;
+ cq->last = 0;
+ cq->is_rx = is_rx;
+ cq->offset_ind = offset_ind;
+
+ /* Alloc doorbell */
+ err = mtnic_alloc_aligned(sizeof(struct mtnic_cq_db_record),
+ (void *)&cq->db, &cq->db_dma, 32);
+ if (err) {
+ DBG("Failed allocating CQ doorbell record\n");
+ return -EADDRINUSE;
+ }
+ memset(cq->db, 0, sizeof(struct mtnic_cq_db_record));
+
+ /* Alloc CQEs buffer */
+ cq->buf_size = size * sizeof(struct mtnic_cqe);
+ err = mtnic_alloc_aligned(cq->buf_size,
+ (void *)&cq->buf, &cq->dma, PAGE_SIZE);
+ if (err) {
+ DBG("Failed allocating CQ buffer\n");
+ free_memblock(cq->db, sizeof(struct mtnic_cq_db_record));
+ return -EADDRINUSE;
+ }
+ memset(cq->buf, 0, cq->buf_size);
+ DBG("Allocated CQ (addr:%p) - size:%x buf:%p buf_size:%x "
+ "dma:%lx db:%p db_dma:%lx\n"
+ "cqn offset:%x \n", cq, cq->size, cq->buf,
+ cq->buf_size, cq->dma, cq->db,
+ cq->db_dma, offset_ind);
+
+
+ /* Set ownership of all CQEs to HW */
+ DBG("Setting HW ownership for CQ:%d\n", num);
+ for (i = 0; i < cq->size; i++) {
+ /* Initial HW ownership is 1 */
+ cq->buf[i].op_tr_own = MTNIC_BIT_CQ_OWN;
+ }
+ return 0;
+}
+
+
+
+/**
+ * mtnic_alloc_resources
+ *
+ * Alloc and configure CQs, Tx, Rx
+ */
+unsigned int
+mtnic_alloc_resources(struct net_device *dev)
+{
+ struct mtnic_port *priv = netdev_priv(dev);
+ int err;
+ int cq_ind = 0;
+ int cq_offset = priv->mtnic->fw.cq_offset;
+
+ /* Alloc 1st CQ */
+ err = mtnic_alloc_cq(dev, cq_ind, &priv->cq[cq_ind], 1 /* RX */,
+ UNITS_BUFFER_SIZE, cq_offset + cq_ind);
+ if (err) {
+ DBG("Failed allocating Rx CQ\n");
+ return -EADDRINUSE;
+ }
+
+
+ /* Alloc RX */
+ err = mtnic_alloc_ring(priv, &priv->rx_ring, UNITS_BUFFER_SIZE,
+ sizeof(struct mtnic_rx_desc), cq_ind, /* RX */1);
+ if (err) {
+ DBG("Failed allocating Rx Ring\n");
+ goto cq0_error;
+ }
+
+
+ ++cq_ind;
+
+ /* alloc 2nd CQ */
+ err = mtnic_alloc_cq(dev, cq_ind, &priv->cq[cq_ind], 0 /* TX */,
+ UNITS_BUFFER_SIZE, cq_offset + cq_ind);
+ if (err) {
+ DBG("Failed allocating Tx CQ\n");
+ goto rx_error;
+ }
+
+ /* Alloc TX */
+ err = mtnic_alloc_ring(priv, &priv->tx_ring, UNITS_BUFFER_SIZE,
+ sizeof(struct mtnic_tx_desc), cq_ind, /* TX */ 0);
+ if (err) {
+ DBG("Failed allocating Tx ring\n");
+ goto cq1_error;
+ }
+
+ return 0;
+
+cq1_error:
+ free_memblock(priv->cq[1].buf, priv->cq[1].buf_size);
+ free_memblock(priv->cq[1].db, sizeof(struct mtnic_cq_db_record));
+
+rx_error:
+ free_memblock(priv->rx_ring.buf, priv->rx_ring.buf_size);
+ free_memblock(priv->rx_ring.db, sizeof(struct mtnic_cq_db_record));
+ mtnic_free_io_buffers(&priv->rx_ring);
+cq0_error:
+ free_memblock(priv->cq[0].buf, priv->cq[0].buf_size);
+ free_memblock(priv->cq[0].db, sizeof(struct mtnic_cq_db_record));
+
+ return -EADDRINUSE;
+}
+
+
+/**
+ * mtnic alloc_eq
+ *
+ * Note: EQ is not used by the driver but must be allocated
+ */
+static int
+mtnic_alloc_eq(struct mtnic *mtnic)
+{
+ int err;
+ unsigned int i;
+ struct mtnic_eqe *eqe_desc = NULL;
+
+ /* Allocating doorbell */
+ mtnic->eq_db = ioremap(mtnic_pci_dev.dev.bar[2] +
+ mtnic->fw.eq_db_offset, sizeof(u32));
+ if (!mtnic->eq_db) {
+ DBG("Couldn't map EQ doorbell, aborting...\n");
+ return -EADDRINUSE;
+ }
+
+ /* Allocating buffer */
+ mtnic->eq.size = NUM_EQES;
+ mtnic->eq.buf_size = mtnic->eq.size * sizeof(struct mtnic_eqe);
+ err = mtnic_alloc_aligned(mtnic->eq.buf_size, (void *)&mtnic->eq.buf,
+ &mtnic->eq.dma, PAGE_SIZE);
+ if (err) {
+ DBG("Failed allocating EQ buffer\n");
+ iounmap(mtnic->eq_db);
+ return -EADDRINUSE;
+ }
+ memset(mtnic->eq.buf, 0, mtnic->eq.buf_size);
+
+ for (i = 0; i < mtnic->eq.size; i++)
+ eqe_desc = mtnic->eq.buf + (sizeof(struct mtnic_eqe) * i);
+ eqe_desc->own |= MTNIC_BIT_EQE_OWN;
+
+ mdelay(20);
+ return 0;
+}
+
+
+
+
+
+
+
+
+
+
+
+/********************************************************************
+*
+* Mtnic commands functions
+* -=-=-=-=-=-=-=-=-=-=-=-=
+*
+*
+*
+*********************************************************************/
+static inline int
+cmdif_go_bit(struct mtnic *mtnic)
+{
+ struct mtnic_if_cmd_reg *hcr = mtnic->hcr;
+ u32 status;
+ int i;
+
+ for (i = 0; i < TBIT_RETRIES; i++) {
+ status = be32_to_cpu(readl(&hcr->status_go_opcode));
+ if ((status & MTNIC_BC_MASK(MTNIC_MASK_CMD_REG_T_BIT)) ==
+ (mtnic->cmd.tbit << MTNIC_BC_OFF(MTNIC_MASK_CMD_REG_T_BIT))) {
+ /* Read expected t-bit - now return go-bit value */
+ return status & MTNIC_BC_MASK(MTNIC_MASK_CMD_REG_GO_BIT);
+ }
+ }
+
+ DBG("Invalid tbit after %d retries!\n", TBIT_RETRIES);
+ return -EBUSY; /* Return busy... */
+}
+
+/* Base Command interface */
+static int
+mtnic_cmd(struct mtnic *mtnic, void *in_imm,
+ void *out_imm, u32 in_modifier, u16 op)
+{
+
+ struct mtnic_if_cmd_reg *hcr = mtnic->hcr;
+ int err = 0;
+ u32 out_param_h = 0;
+ u32 out_param_l = 0;
+ u32 in_param_h = 0;
+ u32 in_param_l = 0;
+
+
+ static u16 token = 0x8000;
+ u32 status;
+ unsigned int timeout = 0;
+
+ token++;
+
+ if ( cmdif_go_bit ( mtnic ) ) {
+ DBG("GO BIT BUSY:%p.\n", hcr + 6);
+ err = -EBUSY;
+ goto out;
+ }
+ if (in_imm) {
+ in_param_h = *((u32*)in_imm);
+ in_param_l = *((u32*)in_imm + 1);
+ } else {
+ in_param_l = cpu_to_be32(mtnic->cmd.mapping);
+ }
+ out_param_l = cpu_to_be32(mtnic->cmd.mapping);
+
+ /* writing to MCR */
+ writel(in_param_h, &hcr->in_param_h);
+ writel(in_param_l, &hcr->in_param_l);
+ writel((u32) cpu_to_be32(in_modifier), &hcr->input_modifier);
+ writel(out_param_h, &hcr->out_param_h);
+ writel(out_param_l, &hcr->out_param_l);
+ writel((u32)cpu_to_be32(token << 16), &hcr->token);
+ wmb();
+
+ /* flip toggle bit before each write to the HCR */
+ mtnic->cmd.tbit = !mtnic->cmd.tbit;
+ writel( ( u32 )
+ cpu_to_be32(MTNIC_BC_MASK(MTNIC_MASK_CMD_REG_GO_BIT) |
+ ( mtnic->cmd.tbit << MTNIC_BC_OFF ( MTNIC_MASK_CMD_REG_T_BIT ) ) | op ),
+ &hcr->status_go_opcode);
+
+ while ( cmdif_go_bit ( mtnic ) && ( timeout <= GO_BIT_TIMEOUT ) ) {
+ mdelay ( 1 );
+ ++timeout;
+ }
+
+ if ( cmdif_go_bit ( mtnic ) ) {
+ DBG("Command opcode:0x%x token:0x%x TIMEOUT.\n", op, token);
+ err = -EBUSY;
+ goto out;
+ }
+
+ if (out_imm) {
+ *((u32 *)out_imm) = readl(&hcr->out_param_h);
+ *((u32 *)out_imm + 1) = readl(&hcr->out_param_l);
+ }
+
+ status = be32_to_cpu((u32)readl(&hcr->status_go_opcode)) >> 24;
+
+ if (status) {
+ DBG("Command opcode:0x%x token:0x%x returned:0x%x\n",
+ op, token, status);
+ return status;
+ }
+
+out:
+ return err;
+}
+
+/* MAP PAGES wrapper */
+static int
+mtnic_map_cmd(struct mtnic *mtnic, u16 op, struct mtnic_pages pages)
+{
+ unsigned int j;
+ u32 addr;
+ unsigned int len;
+ u32 *page_arr = mtnic->cmd.buf;
+ int nent = 0;
+ int err = 0;
+
+ memset(page_arr, 0, PAGE_SIZE);
+
+ len = PAGE_SIZE * pages.num;
+ pages.buf = (u32 *)umalloc(PAGE_SIZE * (pages.num + 1));
+ addr = PAGE_SIZE + ((virt_to_bus(pages.buf) & 0xfffff000) + PAGE_SIZE);
+ DBG("Mapping pages: size: %x address: %p\n", pages.num, pages.buf);
+
+ if (addr & (PAGE_MASK)) {
+ DBG("Got FW area not aligned to %d (%llx/%x)\n",
+ PAGE_SIZE, (u64) addr, len);
+ return -EADDRINUSE;
+ }
+
+ /* Function maps each PAGE seperately */
+ for (j = 0; j < len; j+= PAGE_SIZE) {
+ page_arr[nent * 4 + 3] = cpu_to_be32(addr + j);
+ if (++nent == MTNIC_MAILBOX_SIZE / 16) {
+ err = mtnic_cmd(mtnic, NULL, NULL, nent, op);
+ if (err)
+ return -EIO;
+ nent = 0;
+ }
+ }
+
+ if (nent) {
+ err = mtnic_cmd(mtnic, NULL, NULL, nent, op);
+ }
+ return err;
+}
+
+
+
+/*
+ * Query FW
+ */
+static int
+mtnic_QUERY_FW ( struct mtnic *mtnic )
+{
+ int err;
+ struct mtnic_if_query_fw_out_mbox *cmd = mtnic->cmd.buf;
+
+ err = mtnic_cmd(mtnic, NULL, NULL, 0, MTNIC_IF_CMD_QUERY_FW);
+ if (err)
+ return -EIO;
+
+ /* Get FW and interface versions */
+ mtnic->fw_ver = ((u64) be16_to_cpu(cmd->rev_maj) << 32) |
+ ((u64) be16_to_cpu(cmd->rev_min) << 16) |
+ (u64) be16_to_cpu(cmd->rev_smin);
+ mtnic->fw.ifc_rev = be16_to_cpu(cmd->ifc_rev);
+
+ /* Get offset for internal error reports (debug) */
+ mtnic->fw.err_buf.offset = be64_to_cpu(cmd->err_buf_start);
+ mtnic->fw.err_buf.size = be32_to_cpu(cmd->err_buf_size);
+
+ DBG("Error buf offset is %llx\n", mtnic->fw.err_buf.offset);
+
+ /* Get number of required FW (4k) pages */
+ mtnic->fw.fw_pages.num = be16_to_cpu(cmd->fw_pages);
+
+ return 0;
+}
+
+
+static int
+mtnic_OPEN_NIC(struct mtnic *mtnic)
+{
+ struct mtnic_if_open_nic_in_mbox *open_nic = mtnic->cmd.buf;
+ u32 extra_pages[2] = {0};
+ int err;
+
+ memset(open_nic, 0, sizeof *open_nic);
+
+ /* port 1 */
+ open_nic->log_rx_p1 = 0;
+ open_nic->log_cq_p1 = 1;
+
+ open_nic->log_tx_p1 = 0;
+ open_nic->steer_p1 = MTNIC_IF_STEER_RSS;
+ /* MAC + VLAN - leave reserved */
+
+ /* port 2 */
+ open_nic->log_rx_p2 = 0;
+ open_nic->log_cq_p2 = 1;
+
+ open_nic->log_tx_p2 = 0;
+ open_nic->steer_p2 = MTNIC_IF_STEER_RSS;
+ /* MAC + VLAN - leave reserved */
+
+ err = mtnic_cmd(mtnic, NULL, extra_pages, 0, MTNIC_IF_CMD_OPEN_NIC);
+
+ mtnic->fw.extra_pages.num = be32_to_cpu(*(extra_pages+1));
+ DBG("Extra pages num is %x\n", mtnic->fw.extra_pages.num);
+ return err;
+}
+
+static int
+mtnic_CONFIG_RX(struct mtnic *mtnic)
+{
+ struct mtnic_if_config_rx_in_imm config_rx;
+
+ memset(&config_rx, 0, sizeof config_rx);
+ return mtnic_cmd(mtnic, &config_rx, NULL, 0, MTNIC_IF_CMD_CONFIG_RX);
+}
+
+static int
+mtnic_CONFIG_TX(struct mtnic *mtnic)
+{
+ struct mtnic_if_config_send_in_imm config_tx;
+
+ config_tx.enph_gpf = 0;
+ return mtnic_cmd(mtnic, &config_tx, NULL, 0, MTNIC_IF_CMD_CONFIG_TX);
+}
+
+static int
+mtnic_HEART_BEAT(struct mtnic_port *priv, u32 *link_state)
+{
+ struct mtnic_if_heart_beat_out_imm heart_beat;
+
+ int err;
+ u32 flags;
+ err = mtnic_cmd(priv->mtnic, NULL, &heart_beat, 0, MTNIC_IF_CMD_HEART_BEAT);
+ if (!err) {
+ flags = be32_to_cpu(heart_beat.flags);
+ if (flags & MTNIC_BC_MASK(MTNIC_MASK_HEAR_BEAT_INT_ERROR)) {
+ DBG("Internal error detected\n");
+ return -EIO;
+ }
+ *link_state = flags &
+ ~((u32) MTNIC_BC_MASK(MTNIC_MASK_HEAR_BEAT_INT_ERROR));
+ }
+ return err;
+}
+
+
+/*
+ * Port commands
+ */
+
+static int
+mtnic_SET_PORT_DEFAULT_RING(struct mtnic_port *priv, u8 port, u16 ring)
+{
+ struct mtnic_if_set_port_default_ring_in_imm def_ring;
+
+ memset(&def_ring, 0, sizeof(def_ring));
+ def_ring.ring = ring;
+ return mtnic_cmd(priv->mtnic, &def_ring, NULL, port + 1,
+ MTNIC_IF_CMD_SET_PORT_DEFAULT_RING);
+}
+
+static int
+mtnic_CONFIG_PORT_RSS_STEER(struct mtnic_port *priv, int port)
+{
+ memset(priv->mtnic->cmd.buf, 0, PAGE_SIZE);
+ return mtnic_cmd(priv->mtnic, NULL, NULL, port + 1,
+ MTNIC_IF_CMD_CONFIG_PORT_RSS_STEER);
+}
+
+static int
+mtnic_SET_PORT_RSS_INDIRECTION(struct mtnic_port *priv, int port)
+
+{
+ memset(priv->mtnic->cmd.buf, 0, PAGE_SIZE);
+ return mtnic_cmd(priv->mtnic, NULL, NULL, port + 1,
+ MTNIC_IF_CMD_SET_PORT_RSS_INDIRECTION);
+}
+
+
+/*
+ * Config commands
+ */
+static int
+mtnic_CONFIG_CQ(struct mtnic_port *priv, int port,
+ u16 cq_ind, struct mtnic_cq *cq)
+{
+ struct mtnic_if_config_cq_in_mbox *config_cq = priv->mtnic->cmd.buf;
+
+ memset(config_cq, 0, sizeof *config_cq);
+ config_cq->cq = cq_ind;
+ config_cq->size = fls(UNITS_BUFFER_SIZE - 1);
+ config_cq->offset = ((cq->dma) & (PAGE_MASK)) >> 6;
+ config_cq->db_record_addr_l = cpu_to_be32(cq->db_dma);
+ config_cq->page_address[1] = cpu_to_be32(cq->dma);
+ DBG("config cq address: %x dma_address: %lx"
+ "offset: %d size %d index: %d\n"
+ , config_cq->page_address[1],cq->dma,
+ config_cq->offset, config_cq->size, config_cq->cq );
+
+ return mtnic_cmd(priv->mtnic, NULL, NULL, port + 1,
+ MTNIC_IF_CMD_CONFIG_CQ);
+}
+
+
+static int
+mtnic_CONFIG_TX_RING(struct mtnic_port *priv, u8 port,
+ u16 ring_ind, struct mtnic_ring *ring)
+{
+ struct mtnic_if_config_send_ring_in_mbox *config_tx_ring = priv->mtnic->cmd.buf;
+ memset(config_tx_ring, 0, sizeof *config_tx_ring);
+ config_tx_ring->ring = cpu_to_be16(ring_ind);
+ config_tx_ring->size = fls(UNITS_BUFFER_SIZE - 1);
+ config_tx_ring->cq = cpu_to_be16(ring->cq);
+ config_tx_ring->page_address[1] = cpu_to_be32(ring->dma);
+
+ return mtnic_cmd(priv->mtnic, NULL, NULL, port + 1,
+ MTNIC_IF_CMD_CONFIG_TX_RING);
+}
+
+static int
+mtnic_CONFIG_RX_RING(struct mtnic_port *priv, u8 port,
+ u16 ring_ind, struct mtnic_ring *ring)
+{
+ struct mtnic_if_config_rx_ring_in_mbox *config_rx_ring = priv->mtnic->cmd.buf;
+ memset(config_rx_ring, 0, sizeof *config_rx_ring);
+ config_rx_ring->ring = ring_ind;
+ MTNIC_BC_PUT(config_rx_ring->stride_size, fls(UNITS_BUFFER_SIZE - 1),
+ MTNIC_MASK_CONFIG_RX_RING_SIZE);
+ MTNIC_BC_PUT(config_rx_ring->stride_size, 1,
+ MTNIC_MASK_CONFIG_RX_RING_STRIDE);
+ config_rx_ring->cq = cpu_to_be16(ring->cq);
+ config_rx_ring->db_record_addr_l = cpu_to_be32(ring->db_dma);
+
+ DBG("Config RX ring starting at address:%lx\n", ring->dma);
+
+ config_rx_ring->page_address[1] = cpu_to_be32(ring->dma);
+
+ return mtnic_cmd(priv->mtnic, NULL, NULL, port + 1,
+ MTNIC_IF_CMD_CONFIG_RX_RING);
+}
+
+static int
+mtnic_CONFIG_EQ(struct mtnic *mtnic)
+{
+ struct mtnic_if_config_eq_in_mbox *eq = mtnic->cmd.buf;
+
+ if (mtnic->eq.dma & (PAGE_MASK)) {
+ DBG("misalligned eq buffer:%lx\n",
+ mtnic->eq.dma);
+ return -EADDRINUSE;
+ }
+
+ memset(eq, 0, sizeof *eq);
+ MTNIC_BC_PUT(eq->offset, mtnic->eq.dma >> 6, MTNIC_MASK_CONFIG_EQ_OFFSET);
+ MTNIC_BC_PUT(eq->size, fls(mtnic->eq.size - 1) - 1, MTNIC_MASK_CONFIG_EQ_SIZE);
+ MTNIC_BC_PUT(eq->int_vector, 0, MTNIC_MASK_CONFIG_EQ_INT_VEC);
+ eq->page_address[1] = cpu_to_be32(mtnic->eq.dma);
+
+ return mtnic_cmd(mtnic, NULL, NULL, 0, MTNIC_IF_CMD_CONFIG_EQ);
+}
+
+
+
+
+static int
+mtnic_SET_RX_RING_ADDR(struct mtnic_port *priv, u8 port, u64* mac)
+{
+ struct mtnic_if_set_rx_ring_addr_in_imm ring_addr;
+ u32 modifier = ((u32) port + 1) << 16;
+
+ memset(&ring_addr, 0, sizeof(ring_addr));
+
+ ring_addr.mac_31_0 = cpu_to_be32(*mac & 0xffffffff);
+ ring_addr.mac_47_32 = cpu_to_be16((*mac >> 32) & 0xffff);
+ ring_addr.flags_vlan_id |= cpu_to_be16(
+ MTNIC_BC_MASK(MTNIC_MASK_SET_RX_RING_ADDR_BY_MAC));
+
+ return mtnic_cmd(priv->mtnic, &ring_addr, NULL, modifier, MTNIC_IF_CMD_SET_RX_RING_ADDR);
+}
+
+static int
+mtnic_SET_PORT_STATE(struct mtnic_port *priv, u8 port, u8 state)
+{
+ struct mtnic_if_set_port_state_in_imm port_state;
+
+ port_state.state = state ? cpu_to_be32(
+ MTNIC_BC_MASK(MTNIC_MASK_CONFIG_PORT_STATE)) : 0;
+ port_state.reserved = 0;
+ return mtnic_cmd(priv->mtnic, &port_state, NULL, port + 1,
+ MTNIC_IF_CMD_SET_PORT_STATE);
+}
+
+static int
+mtnic_SET_PORT_MTU(struct mtnic_port *priv, u8 port, u16 mtu)
+{
+ struct mtnic_if_set_port_mtu_in_imm set_mtu;
+
+ memset(&set_mtu, 0, sizeof(set_mtu));
+ set_mtu.mtu = cpu_to_be16(mtu);
+ return mtnic_cmd(priv->mtnic, &set_mtu, NULL, port + 1,
+ MTNIC_IF_CMD_SET_PORT_MTU);
+}
+
+/*
+static int
+mtnic_CONFIG_PORT_VLAN_FILTER(struct mtnic_port *priv, int port)
+{
+ struct mtnic_if_config_port_vlan_filter_in_mbox *vlan_filter = priv->mtnic->cmd.buf;
+
+ // When no vlans are configured we disable the filter
+ // (i.e., pass all vlans) because we ignore them anyhow
+ memset(vlan_filter, 0xff, sizeof(*vlan_filter));
+ return mtnic_cmd(priv->mtnic, NULL, NULL, port + 1,
+ MTNIC_IF_CMD_CONFIG_PORT_VLAN_FILTER);
+}
+*/
+
+
+static int
+mtnic_RELEASE_RESOURCE(struct mtnic_port *priv, u8 port, u8 type, u8 index)
+{
+ struct mtnic_if_release_resource_in_imm rel;
+ memset(&rel, 0, sizeof rel);
+ rel.index = index;
+ rel.type = type;
+ return mtnic_cmd ( priv->mtnic,
+ &rel, NULL, ( type == MTNIC_IF_RESOURCE_TYPE_EQ ) ?
+ 0 : port + 1, MTNIC_IF_CMD_RELEASE_RESOURCE );
+}
+
+
+static int
+mtnic_QUERY_CAP(struct mtnic *mtnic, u8 index, u8 mod, u64 *result)
+{
+ struct mtnic_if_query_cap_in_imm cap;
+ u32 out_imm[2];
+ int err;
+
+ memset(&cap, 0, sizeof cap);
+ cap.cap_index = index;
+ cap.cap_modifier = mod;
+ err = mtnic_cmd(mtnic, &cap, &out_imm, 0, MTNIC_IF_CMD_QUERY_CAP);
+
+ *((u32*)result) = be32_to_cpu(*(out_imm+1));
+ *((u32*)result + 1) = be32_to_cpu(*out_imm);
+
+ DBG("Called Query cap with index:0x%x mod:%d result:0x%llx"
+ " error:%d\n", index, mod, *result, err);
+ return err;
+}
+
+
+#define DO_QUERY_CAP(cap, mod, var) \
+ err = mtnic_QUERY_CAP(mtnic, cap, mod, &result);\
+ if (err) \
+ return err; \
+ (var) = result
+
+static int
+mtnic_query_num_ports(struct mtnic *mtnic)
+{
+ int err = 0;
+ u64 result;
+
+ DO_QUERY_CAP(MTNIC_IF_CAP_NUM_PORTS, 0, mtnic->fw.num_ports);
+
+ return 0;
+}
+
+static int
+mtnic_query_mac(struct mtnic *mtnic)
+{
+ int err = 0;
+ int i;
+ u64 result;
+
+ for (i = 0; i < mtnic->fw.num_ports; i++) {
+ DO_QUERY_CAP(MTNIC_IF_CAP_DEFAULT_MAC, i + 1, mtnic->fw.mac[i]);
+ }
+
+ return 0;
+}
+
+static int
+mtnic_query_offsets(struct mtnic *mtnic)
+{
+ int err;
+ int i;
+ u64 result;
+
+ DO_QUERY_CAP(MTNIC_IF_CAP_MEM_KEY,
+ MTNIC_IF_MEM_TYPE_SNOOP,
+ mtnic->fw.mem_type_snoop_be);
+ mtnic->fw.mem_type_snoop_be = cpu_to_be32(mtnic->fw.mem_type_snoop_be);
+ DO_QUERY_CAP(MTNIC_IF_CAP_TX_CQ_DB_OFFSET, 0, mtnic->fw.txcq_db_offset);
+ DO_QUERY_CAP(MTNIC_IF_CAP_EQ_DB_OFFSET, 0, mtnic->fw.eq_db_offset);
+
+ for (i = 0; i < mtnic->fw.num_ports; i++) {
+ DO_QUERY_CAP(MTNIC_IF_CAP_CQ_OFFSET, i + 1, mtnic->fw.cq_offset);
+ DO_QUERY_CAP(MTNIC_IF_CAP_TX_OFFSET, i + 1, mtnic->fw.tx_offset[i]);
+ DO_QUERY_CAP(MTNIC_IF_CAP_RX_OFFSET, i + 1, mtnic->fw.rx_offset[i]);
+ DBG("--> Port %d CQ offset:0x%x\n", i, mtnic->fw.cq_offset);
+ DBG("--> Port %d Tx offset:0x%x\n", i, mtnic->fw.tx_offset[i]);
+ DBG("--> Port %d Rx offset:0x%x\n", i, mtnic->fw.rx_offset[i]);
+ }
+
+ mdelay(20);
+ return 0;
+}
+
+
+
+
+
+
+
+
+
+
+
+/********************************************************************
+*
+* MTNIC initalization functions
+*
+*
+*
+*
+*********************************************************************/
+
+/**
+ * Reset device
+ */
+void
+mtnic_reset ( void )
+{
+ void *reset = ioremap ( mtnic_pci_dev.dev.bar[0] + MTNIC_RESET_OFFSET,
+ 4 );
+ writel ( cpu_to_be32 ( 1 ), reset );
+ iounmap ( reset );
+}
+
+
+/**
+ * Restore PCI config
+ */
+static int
+restore_config(void)
+{
+ int i;
+ int rc;
+
+ for (i = 0; i < 64; ++i) {
+ if (i != 22 && i != 23) {
+ rc = pci_write_config_dword(mtnic_pci_dev.dev.dev,
+ i << 2,
+ mtnic_pci_dev.dev.
+ dev_config_space[i]);
+ if (rc)
+ return rc;
+ }
+ }
+ return 0;
+}
+
+
+
+/**
+ * Init PCI configuration
+ */
+static int
+mtnic_init_pci(struct pci_device *dev)
+{
+ int i;
+ int err;
+
+ /* save bars */
+ DBG("bus=%d devfn=0x%x\n", dev->bus, dev->devfn);
+ for (i = 0; i < 6; ++i) {
+ mtnic_pci_dev.dev.bar[i] =
+ pci_bar_start(dev, PCI_BASE_ADDRESS_0 + (i << 2));
+ DBG("bar[%d]= 0x%08lx \n", i, mtnic_pci_dev.dev.bar[i]);
+ }
+
+ /* save config space */
+ for (i = 0; i < 64; ++i) {
+ err = pci_read_config_dword(dev, i << 2,
+ &mtnic_pci_dev.dev.
+ dev_config_space[i]);
+ if (err) {
+ DBG("Can not save configuration space");
+ return err;
+ }
+ }
+
+ mtnic_pci_dev.dev.dev = dev;
+
+ return 0;
+}
+
+/**
+ * Initial hardware
+ */
+static inline
+int mtnic_init_card(struct mtnic *mtnic)
+{
+ int err = 0;
+
+
+ /* Alloc command interface */
+ err = mtnic_alloc_cmdif ( mtnic );
+ if (err) {
+ DBG("Failed to init command interface, aborting\n");
+ return -EADDRINUSE;
+ }
+
+
+ /**
+ * Bring up HW
+ */
+ err = mtnic_QUERY_FW ( mtnic );
+ if (err) {
+ DBG("QUERY_FW command failed, aborting\n");
+ goto cmd_error;
+ }
+ DBG("Command interface revision:%d\n", mtnic->fw.ifc_rev);
+
+ /* Allocate memory for FW and start it */
+ err = mtnic_map_cmd(mtnic, MTNIC_IF_CMD_MAP_FW, mtnic->fw.fw_pages);
+ if (err) {
+ DBG("Eror In MAP_FW\n");
+ if (mtnic->fw.fw_pages.buf)
+ ufree((intptr_t)mtnic->fw.fw_pages.buf);
+ goto cmd_error;
+ }
+
+ /* Run firmware */
+ err = mtnic_cmd(mtnic, NULL, NULL, 0, MTNIC_IF_CMD_RUN_FW);
+ if (err) {
+ DBG("Eror In RUN FW\n");
+ goto map_fw_error;
+ }
+
+ DBG("FW version:%d.%d.%d\n",
+ (u16) (mtnic->fw_ver >> 32),
+ (u16) ((mtnic->fw_ver >> 16) & 0xffff),
+ (u16) (mtnic->fw_ver & 0xffff));
+
+
+ /* Query num ports */
+ err = mtnic_query_num_ports(mtnic);
+ if (err) {
+ DBG("Insufficient resources, aborting\n");
+ goto map_fw_error;
+ }
+
+ /* Open NIC */
+ err = mtnic_OPEN_NIC(mtnic);
+ if (err) {
+ DBG("Failed opening NIC, aborting\n");
+ goto map_fw_error;
+ }
+
+ /* Allocate and map pages worksace */
+ err = mtnic_map_cmd(mtnic, MTNIC_IF_CMD_MAP_PAGES, mtnic->fw.extra_pages);
+ if (err) {
+ DBG("Couldn't allocate %x FW extra pages, aborting\n",
+ mtnic->fw.extra_pages.num);
+ if (mtnic->fw.extra_pages.buf)
+ ufree((intptr_t)mtnic->fw.extra_pages.buf);
+ goto map_fw_error;
+ }
+
+
+ /* Get device information */
+ err = mtnic_query_mac(mtnic);
+ if (err) {
+ DBG("Insufficient resources in quesry mac, aborting\n");
+ goto map_fw_error;
+ }
+
+ /* Get device offsets */
+ err = mtnic_query_offsets(mtnic);
+ if (err) {
+ DBG("Failed retrieving resource offests, aborting\n");
+ ufree((intptr_t)mtnic->fw.extra_pages.buf);
+ goto map_extra_error;
+ }
+
+
+ /* Alloc EQ */
+ err = mtnic_alloc_eq(mtnic);
+ if (err) {
+ DBG("Failed init shared resources. error: %d\n", err);
+ goto map_extra_error;
+ }
+
+ /* Configure HW */
+ err = mtnic_CONFIG_EQ(mtnic);
+ if (err) {
+ DBG("Failed configuring EQ\n");
+ goto eq_error;
+ }
+ err = mtnic_CONFIG_RX(mtnic);
+ if (err) {
+ DBG("Failed Rx configuration\n");
+ goto eq_error;
+ }
+ err = mtnic_CONFIG_TX(mtnic);
+ if (err) {
+ DBG("Failed Tx configuration\n");
+ goto eq_error;
+ }
+
+
+ return 0;
+
+
+eq_error:
+ iounmap(mtnic->eq_db);
+ free_memblock(mtnic->eq.buf, mtnic->eq.buf_size);
+map_extra_error:
+ ufree((intptr_t)mtnic->fw.extra_pages.buf);
+map_fw_error:
+ ufree((intptr_t)mtnic->fw.fw_pages.buf);
+
+cmd_error:
+ iounmap(mtnic->hcr);
+ free_memblock(mtnic->cmd.buf, PAGE_SIZE);
+
+ return -EADDRINUSE;
+}
+
+
+
+
+
+
+
+
+
+
+/*******************************************************************
+*
+* Process functions
+*
+* process compliations of TX and RX
+*
+*
+********************************************************************/
+void mtnic_process_tx_cq(struct mtnic_port *priv, struct net_device *dev,
+ struct mtnic_cq *cq)
+{
+ struct mtnic_cqe *cqe = cq->buf;
+ struct mtnic_ring *ring = &priv->tx_ring;
+ u16 index;
+
+
+ index = cq->last & (cq->size-1);
+ cqe = &cq->buf[index];
+
+ /* Owner bit changes every round */
+ while (XNOR(cqe->op_tr_own & MTNIC_BIT_CQ_OWN, cq->last & cq->size)) {
+ netdev_tx_complete (dev, ring->iobuf[index]);
+ ++cq->last;
+ index = cq->last & (cq->size-1);
+ cqe = &cq->buf[index];
+ }
+
+ /* Update consumer index */
+ cq->db->update_ci = cpu_to_be32(cq->last & 0xffffff);
+ wmb(); /* ensure HW sees CQ consumer before we post new buffers */
+ ring->cons = cq->last;
+}
+
+
+int mtnic_process_rx_cq(struct mtnic_port *priv,
+ struct net_device *dev,
+ struct mtnic_cq *cq)
+{
+ struct mtnic_cqe *cqe;
+ struct mtnic_ring *ring = &priv->rx_ring;
+ int index;
+ int err;
+ struct io_buffer *rx_iob;
+ unsigned int length;
+
+
+ /* We assume a 1:1 mapping between CQEs and Rx descriptors, so Rx
+ * descriptor offset can be deduced from the CQE index instead of
+ * reading 'cqe->index' */
+ index = cq->last & (cq->size-1);
+ cqe = &cq->buf[index];
+
+ /* Process all completed CQEs */
+ while (XNOR(cqe->op_tr_own & MTNIC_BIT_CQ_OWN, cq->last & cq->size)) {
+ /* Drop packet on bad receive or bad checksum */
+ if ((cqe->op_tr_own & 0x1f) == MTNIC_OPCODE_ERROR) {
+ DBG("CQE completed with error - vendor \n");
+ free_iob(ring->iobuf[index]);
+ goto next;
+ }
+ if (cqe->enc_bf & MTNIC_BIT_BAD_FCS) {
+ DBG("Accepted packet with bad FCS\n");
+ free_iob(ring->iobuf[index]);
+ goto next;
+ }
+
+ /*
+ * Packet is OK - process it.
+ */
+ length = be32_to_cpu(cqe->byte_cnt);
+ rx_iob = ring->iobuf[index];
+ iob_put(rx_iob, length);
+
+ /* Add this packet to the receive queue. */
+ netdev_rx(dev, rx_iob);
+ ring->iobuf[index] = NULL;
+
+next:
+ ++cq->last;
+ index = cq->last & (cq->size-1);
+ cqe = &cq->buf[index];
+
+
+
+ }
+
+ /* Update consumer index */
+ cq->db->update_ci = cpu_to_be32(cq->last & 0xffffff);
+ wmb(); /* ensure HW sees CQ consumer before we post new buffers */
+ ring->cons = cq->last;
+
+ if (ring->prod - ring->cons < (MAX_GAP_PROD_CONS)) {
+ err = mtnic_alloc_iobuf(priv, &priv->rx_ring, DEF_IOBUF_SIZE);
+ if (err) {
+ DBG("ERROR Allocating io buffer");
+ return -EADDRINUSE;
+ }
+ }
+
+ return 0;
+}
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+/********************************************************************
+*
+* net_device functions
+*
+*
+* open, poll, close, probe, disable, irq
+*
+*********************************************************************/
+static int
+mtnic_open(struct net_device *dev)
+{
+ struct mtnic_port *priv = netdev_priv(dev);
+
+ int err = 0;
+ struct mtnic_ring *ring;
+ struct mtnic_cq *cq;
+ int cq_ind = 0;
+ u32 dev_link_state;
+ int link_check;
+
+ DBG("starting port:%d, MAC Address: 0x%12llx\n",
+ priv->port, priv->mtnic->fw.mac[priv->port]);
+
+ /* Alloc and configure CQs, TX, RX */
+ err = mtnic_alloc_resources ( dev );
+ if (err) {
+ DBG("Error allocating resources\n");
+ return -EADDRINUSE;
+ }
+
+ /* Pass CQs configuration to HW */
+ for (cq_ind = 0; cq_ind < NUM_CQS; ++cq_ind) {
+ cq = &priv->cq[cq_ind];
+ err = mtnic_CONFIG_CQ(priv, priv->port, cq_ind, cq);
+ if (err) {
+ DBG("Failed configuring CQ:%d error %d\n",
+ cq_ind, err);
+ if (cq_ind)
+ goto cq_error;
+ else
+ goto allocation_error;
+ }
+ /* Update consumer index */
+ cq->db->update_ci = cpu_to_be32(cq->last & 0xffffff);
+ }
+
+
+
+ /* Pass Tx configuration to HW */
+ ring = &priv->tx_ring;
+ err = mtnic_CONFIG_TX_RING(priv, priv->port, 0, ring);
+ if (err) {
+ DBG("Failed configuring Tx ring:0\n");
+ goto cq_error;
+ }
+
+ /* Pass RX configuration to HW */
+ ring = &priv->rx_ring;
+ err = mtnic_CONFIG_RX_RING(priv, priv->port, 0, ring);
+ if (err) {
+ DBG("Failed configuring Rx ring:0\n");
+ goto tx_error;
+ }
+
+ /* Configure Rx steering */
+ err = mtnic_CONFIG_PORT_RSS_STEER(priv, priv->port);
+ if (!err)
+ err = mtnic_SET_PORT_RSS_INDIRECTION(priv, priv->port);
+ if (err) {
+ DBG("Failed configuring RSS steering\n");
+ goto rx_error;
+ }
+
+
+ /* Set the port default ring to ring 0 */
+ err = mtnic_SET_PORT_DEFAULT_RING(priv, priv->port, 0);
+ if (err) {
+ DBG("Failed setting default ring\n");
+ goto rx_error;
+ }
+
+ /* Set Mac address */
+ err = mtnic_SET_RX_RING_ADDR(priv, priv->port, &priv->mtnic->fw.mac[priv->port]);
+ if (err) {
+ DBG("Failed setting default MAC address\n");
+ goto rx_error;
+ }
+
+ /* Set MTU */
+ err = mtnic_SET_PORT_MTU(priv, priv->port, DEF_MTU);
+ if (err) {
+ DBG("Failed setting MTU\n");
+ goto rx_error;
+ }
+
+ /* Configure VLAN filter */
+ /* By adding this function, The second port won't accept packets
+ err = mtnic_CONFIG_PORT_VLAN_FILTER(priv, priv->port);
+ if (err) {
+ DBG("Failed configuring VLAN filter\n");
+ goto rx_error;
+ }
+ */
+
+
+ /* Bring up physical link */
+ err = mtnic_SET_PORT_STATE(priv, priv->port, 1);
+ if (err) {
+ DBG("Failed bringing up port\n");
+ goto rx_error;
+ }
+
+ /* PORT IS UP */
+ priv->state = CARD_UP;
+
+
+ /* Checking Link is up */
+ DBG ( "Checking if link is up\n" );
+
+
+ for ( link_check = 0; link_check < CHECK_LINK_TIMES; link_check ++ ) {
+ /* Let link state stabilize if cable was connected */
+ mdelay ( DELAY_LINK_CHECK );
+
+ err = mtnic_HEART_BEAT(priv, &dev_link_state);
+ if (err) {
+ DBG("Failed getting device link state\n");
+ return -ENETDOWN;
+ }
+
+ if ( dev_link_state & priv->port ) {
+ /* Link is up */
+ break;
+ }
+ }
+
+
+ if ( ! ( dev_link_state & 0x3 ) ) {
+ DBG("Link down, check cables and restart\n");
+ netdev_link_down ( dev );
+ return -ENETDOWN;
+ }
+
+ DBG ( "Link is up!\n" );
+
+ /* Mark as link up */
+ netdev_link_up ( dev );
+
+ return 0;
+
+rx_error:
+ err = mtnic_RELEASE_RESOURCE(priv, priv->port,
+ MTNIC_IF_RESOURCE_TYPE_RX_RING, 0);
+tx_error:
+ err |= mtnic_RELEASE_RESOURCE(priv, priv->port,
+ MTNIC_IF_RESOURCE_TYPE_TX_RING, 0);
+
+cq_error:
+ while (cq_ind) {
+ err |= mtnic_RELEASE_RESOURCE(priv, priv->port,
+ MTNIC_IF_RESOURCE_TYPE_CQ, --cq_ind);
+ }
+ if (err)
+ DBG("Eror Releasing resources\n");
+
+allocation_error:
+
+ free_memblock(priv->tx_ring.buf, priv->tx_ring.buf_size);
+ iounmap(priv->tx_ring.txcq_db);
+ free_memblock(priv->cq[1].buf, priv->cq[1].buf_size);
+ free_memblock(priv->cq[1].db, sizeof(struct mtnic_cq_db_record));
+ free_memblock(priv->rx_ring.buf, priv->rx_ring.buf_size);
+ free_memblock(priv->rx_ring.db, sizeof(struct mtnic_cq_db_record));
+ free_memblock(priv->cq[0].buf, priv->cq[0].buf_size);
+ free_memblock(priv->cq[0].db, sizeof(struct mtnic_cq_db_record));
+
+ mtnic_free_io_buffers(&priv->rx_ring);
+
+ return -ENETDOWN;
+}
+
+
+
+
+/** Check if we got completion for receive and transmit and
+ * check the line with heart_bit command */
+static void
+mtnic_poll ( struct net_device *dev )
+{
+ struct mtnic_port *priv = netdev_priv(dev);
+ struct mtnic_cq *cq;
+ u32 dev_link_state;
+ int err;
+ unsigned int i;
+
+ /* In case of an old error then return */
+ if (priv->state != CARD_UP)
+ return;
+
+ /* We do not check the device every call _poll call,
+ since it will slow it down */
+ if ((priv->poll_counter % ROUND_TO_CHECK) == 0) {
+ /* Check device */
+ err = mtnic_HEART_BEAT(priv, &dev_link_state);
+ if (err) {
+ DBG("Device has internal error\n");
+ priv->state = CARD_LINK_DOWN;
+ return;
+ }
+ if (!(dev_link_state & 0x3)) {
+ DBG("Link down, check cables and restart\n");
+ priv->state = CARD_LINK_DOWN;
+ return;
+ }
+ }
+ /* Polling CQ */
+ for (i = 0; i < NUM_CQS; i++) {
+ cq = &priv->cq[i]; //Passing on the 2 cqs.
+
+ if (cq->is_rx) {
+ err = mtnic_process_rx_cq(priv, cq->dev, cq);
+ if (err) {
+ priv->state = CARD_LINK_DOWN;
+ DBG(" Error allocating RX buffers\n");
+ return;
+ }
+ } else {
+ mtnic_process_tx_cq(priv, cq->dev, cq);
+ }
+ }
+ ++ priv->poll_counter;
+}
+
+
+
+static int
+mtnic_transmit( struct net_device *dev, struct io_buffer *iobuf )
+{
+
+ struct mtnic_port *priv = netdev_priv(dev);
+ struct mtnic_ring *ring;
+ struct mtnic_tx_desc *tx_desc;
+ struct mtnic_data_seg *data;
+ u32 index;
+
+ /* In case of an error then return */
+ if (priv->state != CARD_UP)
+ return -ENETDOWN;
+
+ ring = &priv->tx_ring;
+
+ index = ring->prod & ring->size_mask;
+ if ((ring->prod - ring->cons) >= ring->size) {
+ DBG("No space left for descriptors!!! cons: %x prod: %x\n",
+ ring->cons, ring->prod);
+ mdelay(5);
+ return -EAGAIN;/* no space left */
+ }
+
+ /* get current descriptor */
+ tx_desc = ring->buf + (index * sizeof(struct mtnic_tx_desc));
+
+ /* Prepare Data Seg */
+ data = &tx_desc->data;
+ data->addr_l = cpu_to_be32((u32)virt_to_bus(iobuf->data));
+ data->count = cpu_to_be32(iob_len(iobuf));
+ data->mem_type = priv->mtnic->fw.mem_type_snoop_be;
+
+ /* Prepare ctrl segement */
+ tx_desc->ctrl.size_vlan = cpu_to_be32(2);
+ tx_desc->ctrl.flags = cpu_to_be32(MTNIC_BIT_TX_COMP |
+ MTNIC_BIT_NO_ICRC);
+ tx_desc->ctrl.op_own = cpu_to_be32(MTNIC_OPCODE_SEND) |
+ ((ring->prod & ring->size) ?
+ cpu_to_be32(MTNIC_BIT_DESC_OWN) : 0);
+
+ /* Attach io_buffer */
+ ring->iobuf[index] = iobuf;
+
+ /* Update producer index */
+ ++ring->prod;
+
+ /* Ring doorbell! */
+ wmb();
+ writel((u32) ring->db_offset, &ring->txcq_db->send_db);
+
+ return 0;
+}
+
+
+static void
+mtnic_close(struct net_device *dev)
+{
+ struct mtnic_port *priv = netdev_priv(dev);
+ int err = 0;
+ DBG("Close called for port:%d\n", priv->port);
+
+ if ( ( priv->state == CARD_UP ) ||
+ ( priv->state == CARD_LINK_DOWN ) ) {
+
+ /* Disable port */
+ err |= mtnic_SET_PORT_STATE(priv, priv->port, 0);
+ /*
+ * Stop HW associated with this port
+ */
+ mdelay(5);
+
+ /* Stop RX */
+ err |= mtnic_RELEASE_RESOURCE(priv, priv->port,
+ MTNIC_IF_RESOURCE_TYPE_RX_RING, 0);
+
+ /* Stop TX */
+ err |= mtnic_RELEASE_RESOURCE(priv, priv->port,
+ MTNIC_IF_RESOURCE_TYPE_TX_RING, 0);
+
+ /* Stop CQs */
+ err |= mtnic_RELEASE_RESOURCE(priv, priv->port,
+ MTNIC_IF_RESOURCE_TYPE_CQ, 0);
+ err |= mtnic_RELEASE_RESOURCE(priv, priv->port,
+ MTNIC_IF_RESOURCE_TYPE_CQ, 1);
+ if (err) {
+ DBG("Close reported error %d\n", err);
+ }
+
+ mdelay ( 10 );
+
+ /* free memory */
+ free_memblock(priv->tx_ring.buf, priv->tx_ring.buf_size);
+ iounmap(priv->tx_ring.txcq_db);
+ free_memblock(priv->cq[1].buf, priv->cq[1].buf_size);
+ free_memblock(priv->cq[1].db, sizeof(struct mtnic_cq_db_record));
+ free_memblock(priv->rx_ring.buf, priv->rx_ring.buf_size);
+ free_memblock(priv->rx_ring.db, sizeof(struct mtnic_cq_db_record));
+ free_memblock(priv->cq[0].buf, priv->cq[0].buf_size);
+ free_memblock(priv->cq[0].db, sizeof(struct mtnic_cq_db_record));
+
+ /* Free RX buffers */
+ mtnic_free_io_buffers(&priv->rx_ring);
+
+
+
+ }
+
+ priv->state = CARD_INITIALIZED;
+
+}
+
+
+static void
+mtnic_disable(struct pci_device *pci)
+{
+
+ int err;
+ int i;
+ struct mtnic *mtnic = pci_get_drvdata(pci);
+
+
+ struct net_device *dev;
+ struct mtnic_port *priv;
+
+ for ( i = ( mtnic->fw.num_ports - 1 ); i >= 0; i-- ) {
+
+ dev = mtnic->netdev[i];
+
+ priv = netdev_priv(dev);
+
+ /* Just in case */
+ if ( ( priv->state == CARD_UP ) ||
+ ( priv->state == CARD_LINK_DOWN ) )
+ mtnic_close ( dev );
+ }
+
+ /* Releasing EQ */
+ priv = netdev_priv ( mtnic->netdev[0] );
+ err = mtnic_RELEASE_RESOURCE(priv, 1,
+ MTNIC_IF_RESOURCE_TYPE_EQ, 0);
+
+ DBG("Calling MTNIC_CLOSE command\n");
+ err |= mtnic_cmd(mtnic, NULL, NULL, 0,
+ MTNIC_IF_CMD_CLOSE_NIC);
+ if (err) {
+ DBG("Error Releasing resources %d\n", err);
+ }
+
+ free_memblock(mtnic->cmd.buf, PAGE_SIZE);
+ iounmap(mtnic->hcr);
+ ufree((intptr_t)mtnic->fw.fw_pages.buf);
+ ufree((intptr_t)mtnic->fw.extra_pages.buf);
+ free_memblock(mtnic->eq.buf, mtnic->eq.buf_size);
+ iounmap(mtnic->eq_db);
+
+
+ for ( i = ( mtnic->fw.num_ports - 1 ); i >= 0; i-- ) {
+ dev = mtnic->netdev[i];
+ unregister_netdev ( dev );
+ netdev_nullify ( dev );
+ netdev_put ( dev );
+ }
+
+ free ( mtnic );
+
+
+ mtnic_reset ();
+ mdelay ( 1000 );
+ /* Restore config, if we would like to retry booting */
+ restore_config ();
+
+
+}
+
+
+
+static void
+mtnic_irq(struct net_device *netdev __unused, int enable __unused)
+{
+ /* Not implemented */
+}
+
+
+
+/** mtnic net device operations */
+static struct net_device_operations mtnic_operations = {
+ .open = mtnic_open,
+ .close = mtnic_close,
+ .transmit = mtnic_transmit,
+ .poll = mtnic_poll,
+ .irq = mtnic_irq,
+};
+
+
+
+
+
+
+
+static int
+mtnic_probe(struct pci_device *pci,
+ const struct pci_device_id *id __unused)
+{
+ struct mtnic_port *priv;
+ struct mtnic *mtnic;
+ int err;
+ u64 mac;
+ int port_index;
+
+
+ adjust_pci_device(pci);
+
+ err = mtnic_init_pci(pci);
+ if (err) {
+ DBG("Error in pci_init\n");
+ return -EIO;
+ }
+
+ mtnic_reset();
+ mdelay(1000);
+
+ err = restore_config();
+ if (err) {
+ DBG("Error in restoring config\n");
+ return err;
+ }
+
+ mtnic = zalloc ( sizeof ( *mtnic ) );
+ if ( ! mtnic ) {
+ DBG ( "Error Allocating mtnic buffer\n" );
+ return -EADDRINUSE;
+ }
+
+ pci_set_drvdata(pci, mtnic);
+
+ mtnic->pdev = pci;
+
+
+ /* Initialize hardware */
+ err = mtnic_init_card ( mtnic );
+ if (err) {
+ DBG("Error in init_card\n");
+ goto err_init_card;
+ }
+
+ for ( port_index = 0; port_index < mtnic->fw.num_ports; port_index ++ ) {
+ /* Initializing net device */
+ mtnic->netdev[port_index] = alloc_etherdev( sizeof ( struct mtnic_port ) );
+ if ( mtnic->netdev[port_index] == NULL ) {
+ DBG("Net device allocation failed\n");
+ goto err_alloc_mtnic;
+ }
+
+ /*
+ * Initialize driver private data
+ */
+
+ mtnic->netdev[port_index]->dev = &pci->dev;
+ priv = netdev_priv ( mtnic->netdev[port_index] );
+ memset ( priv, 0, sizeof ( struct mtnic_port ) );
+ priv->mtnic = mtnic;
+ priv->netdev = mtnic->netdev[port_index];
+
+ /* Attach pci device */
+ netdev_init(mtnic->netdev[port_index], &mtnic_operations);
+
+ /* Set port number */
+ priv->port = port_index;
+
+ /* Set state */
+ priv->state = CARD_DOWN;
+ }
+
+
+ int mac_idx;
+ for ( port_index = 0; port_index < mtnic->fw.num_ports; port_index ++ ) {
+ priv = netdev_priv ( mtnic->netdev[port_index] );
+ /* Program the MAC address */
+ mac = priv->mtnic->fw.mac[port_index];
+ for (mac_idx = 0; mac_idx < MAC_ADDRESS_SIZE; ++mac_idx) {
+ mtnic->netdev[port_index]->hw_addr[MAC_ADDRESS_SIZE - mac_idx - 1] = mac & 0xFF;
+ mac = mac >> 8;
+ }
+
+ if ( register_netdev ( mtnic->netdev[port_index] ) ) {
+ DBG("Netdev registration failed\n");
+ priv->state = CARD_INITIALIZED;
+ goto err_alloc_mtnic;
+ }
+ }
+
+
+ return 0;
+
+err_alloc_mtnic:
+ free ( mtnic );
+err_init_card:
+ return -EIO;
+}
+
+
+
+
+static struct pci_device_id mtnic_nics[] = {
+ PCI_ROM ( 0x15b3, 0x6368, "mt25448", "Mellanox ConnectX EN driver", 0 ),
+ PCI_ROM ( 0x15b3, 0x6372, "mt25458", "Mellanox ConnectX ENt driver", 0 ),
+ PCI_ROM ( 0x15b3, 0x6750, "mt26448", "Mellanox ConnectX EN GEN2 driver", 0 ),
+ PCI_ROM ( 0x15b3, 0x675a, "mt26458", "Mellanox ConnectX ENt GEN2 driver", 0 ),
+};
+
+struct pci_driver mtnic_driver __pci_driver = {
+ .ids = mtnic_nics,
+ .id_count = sizeof(mtnic_nics) / sizeof(mtnic_nics[0]),
+ .probe = mtnic_probe,
+ .remove = mtnic_disable,
+};
+