Commit 453a6082 authored by Yevgeny Petrilin's avatar Yevgeny Petrilin Committed by David S. Miller

mlx4_en: Giving interface name in debug messages

For each debug message, the message will show interface name in case
that the net device was registered, and PCI bus ID with port number
if we were not registered yet. Messages that are not port/netdev specific
stayed in the old format
Signed-off-by: default avatarYevgeny Petrilin <yevgenyp@mellanox.co.il>
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parent f771bef9
This diff is collapsed.
...@@ -371,7 +371,7 @@ static int mlx4_en_set_pauseparam(struct net_device *dev, ...@@ -371,7 +371,7 @@ static int mlx4_en_set_pauseparam(struct net_device *dev,
priv->prof->rx_pause, priv->prof->rx_pause,
priv->prof->rx_ppp); priv->prof->rx_ppp);
if (err) if (err)
mlx4_err(mdev, "Failed setting pause params to\n"); en_err(priv, "Failed setting pause params\n");
return err; return err;
} }
...@@ -421,13 +421,13 @@ static int mlx4_en_set_ringparam(struct net_device *dev, ...@@ -421,13 +421,13 @@ static int mlx4_en_set_ringparam(struct net_device *dev,
err = mlx4_en_alloc_resources(priv); err = mlx4_en_alloc_resources(priv);
if (err) { if (err) {
mlx4_err(mdev, "Failed reallocating port resources\n"); en_err(priv, "Failed reallocating port resources\n");
goto out; goto out;
} }
if (port_up) { if (port_up) {
err = mlx4_en_start_port(dev); err = mlx4_en_start_port(dev);
if (err) if (err)
mlx4_err(mdev, "Failed starting port\n"); en_err(priv, "Failed starting port\n");
} }
out: out:
......
...@@ -114,7 +114,7 @@ static int mlx4_en_init_allocator(struct mlx4_en_priv *priv, ...@@ -114,7 +114,7 @@ static int mlx4_en_init_allocator(struct mlx4_en_priv *priv,
goto out; goto out;
page_alloc->offset = priv->frag_info[i].frag_align; page_alloc->offset = priv->frag_info[i].frag_align;
mlx4_dbg(DRV, priv, "Initialized allocator:%d with page:%p\n", en_dbg(DRV, priv, "Initialized allocator:%d with page:%p\n",
i, page_alloc->page); i, page_alloc->page);
} }
return 0; return 0;
...@@ -136,7 +136,7 @@ static void mlx4_en_destroy_allocator(struct mlx4_en_priv *priv, ...@@ -136,7 +136,7 @@ static void mlx4_en_destroy_allocator(struct mlx4_en_priv *priv,
for (i = 0; i < priv->num_frags; i++) { for (i = 0; i < priv->num_frags; i++) {
page_alloc = &ring->page_alloc[i]; page_alloc = &ring->page_alloc[i];
mlx4_dbg(DRV, priv, "Freeing allocator:%d count:%d\n", en_dbg(DRV, priv, "Freeing allocator:%d count:%d\n",
i, page_count(page_alloc->page)); i, page_count(page_alloc->page));
put_page(page_alloc->page); put_page(page_alloc->page);
...@@ -214,10 +214,10 @@ static void mlx4_en_free_rx_desc(struct mlx4_en_priv *priv, ...@@ -214,10 +214,10 @@ static void mlx4_en_free_rx_desc(struct mlx4_en_priv *priv,
skb_frags = ring->rx_info + (index << priv->log_rx_info); skb_frags = ring->rx_info + (index << priv->log_rx_info);
for (nr = 0; nr < priv->num_frags; nr++) { for (nr = 0; nr < priv->num_frags; nr++) {
mlx4_dbg(DRV, priv, "Freeing fragment:%d\n", nr); en_dbg(DRV, priv, "Freeing fragment:%d\n", nr);
dma = be64_to_cpu(rx_desc->data[nr].addr); dma = be64_to_cpu(rx_desc->data[nr].addr);
mlx4_dbg(DRV, priv, "Unmaping buffer at dma:0x%llx\n", (u64) dma); en_dbg(DRV, priv, "Unmaping buffer at dma:0x%llx\n", (u64) dma);
pci_unmap_single(mdev->pdev, dma, skb_frags[nr].size, pci_unmap_single(mdev->pdev, dma, skb_frags[nr].size,
PCI_DMA_FROMDEVICE); PCI_DMA_FROMDEVICE);
put_page(skb_frags[nr].page); put_page(skb_frags[nr].page);
...@@ -226,7 +226,6 @@ static void mlx4_en_free_rx_desc(struct mlx4_en_priv *priv, ...@@ -226,7 +226,6 @@ static void mlx4_en_free_rx_desc(struct mlx4_en_priv *priv,
static int mlx4_en_fill_rx_buffers(struct mlx4_en_priv *priv) static int mlx4_en_fill_rx_buffers(struct mlx4_en_priv *priv)
{ {
struct mlx4_en_dev *mdev = priv->mdev;
struct mlx4_en_rx_ring *ring; struct mlx4_en_rx_ring *ring;
int ring_ind; int ring_ind;
int buf_ind; int buf_ind;
...@@ -239,12 +238,12 @@ static int mlx4_en_fill_rx_buffers(struct mlx4_en_priv *priv) ...@@ -239,12 +238,12 @@ static int mlx4_en_fill_rx_buffers(struct mlx4_en_priv *priv)
if (mlx4_en_prepare_rx_desc(priv, ring, if (mlx4_en_prepare_rx_desc(priv, ring,
ring->actual_size)) { ring->actual_size)) {
if (ring->actual_size < MLX4_EN_MIN_RX_SIZE) { if (ring->actual_size < MLX4_EN_MIN_RX_SIZE) {
mlx4_err(mdev, "Failed to allocate " en_err(priv, "Failed to allocate "
"enough rx buffers\n"); "enough rx buffers\n");
return -ENOMEM; return -ENOMEM;
} else { } else {
new_size = rounddown_pow_of_two(ring->actual_size); new_size = rounddown_pow_of_two(ring->actual_size);
mlx4_warn(mdev, "Only %d buffers allocated " en_warn(priv, "Only %d buffers allocated "
"reducing ring size to %d", "reducing ring size to %d",
ring->actual_size, new_size); ring->actual_size, new_size);
goto reduce_rings; goto reduce_rings;
...@@ -282,8 +281,7 @@ static int mlx4_en_fill_rx_buf(struct net_device *dev, ...@@ -282,8 +281,7 @@ static int mlx4_en_fill_rx_buf(struct net_device *dev,
ring->size_mask); ring->size_mask);
if (err) { if (err) {
if (netif_msg_rx_err(priv)) if (netif_msg_rx_err(priv))
mlx4_warn(priv->mdev, en_warn(priv, "Failed preparing rx descriptor\n");
"Failed preparing rx descriptor\n");
priv->port_stats.rx_alloc_failed++; priv->port_stats.rx_alloc_failed++;
break; break;
} }
...@@ -301,14 +299,14 @@ static void mlx4_en_free_rx_buf(struct mlx4_en_priv *priv, ...@@ -301,14 +299,14 @@ static void mlx4_en_free_rx_buf(struct mlx4_en_priv *priv,
{ {
int index; int index;
mlx4_dbg(DRV, priv, "Freeing Rx buf - cons:%d prod:%d\n", en_dbg(DRV, priv, "Freeing Rx buf - cons:%d prod:%d\n",
ring->cons, ring->prod); ring->cons, ring->prod);
/* Unmap and free Rx buffers */ /* Unmap and free Rx buffers */
BUG_ON((u32) (ring->prod - ring->cons) > ring->actual_size); BUG_ON((u32) (ring->prod - ring->cons) > ring->actual_size);
while (ring->cons != ring->prod) { while (ring->cons != ring->prod) {
index = ring->cons & ring->size_mask; index = ring->cons & ring->size_mask;
mlx4_dbg(DRV, priv, "Processing descriptor:%d\n", index); en_dbg(DRV, priv, "Processing descriptor:%d\n", index);
mlx4_en_free_rx_desc(priv, ring, index); mlx4_en_free_rx_desc(priv, ring, index);
++ring->cons; ++ring->cons;
} }
...@@ -373,10 +371,10 @@ int mlx4_en_create_rx_ring(struct mlx4_en_priv *priv, ...@@ -373,10 +371,10 @@ int mlx4_en_create_rx_ring(struct mlx4_en_priv *priv,
sizeof(struct skb_frag_struct)); sizeof(struct skb_frag_struct));
ring->rx_info = vmalloc(tmp); ring->rx_info = vmalloc(tmp);
if (!ring->rx_info) { if (!ring->rx_info) {
mlx4_err(mdev, "Failed allocating rx_info ring\n"); en_err(priv, "Failed allocating rx_info ring\n");
return -ENOMEM; return -ENOMEM;
} }
mlx4_dbg(DRV, priv, "Allocated rx_info ring at addr:%p size:%d\n", en_dbg(DRV, priv, "Allocated rx_info ring at addr:%p size:%d\n",
ring->rx_info, tmp); ring->rx_info, tmp);
err = mlx4_alloc_hwq_res(mdev->dev, &ring->wqres, err = mlx4_alloc_hwq_res(mdev->dev, &ring->wqres,
...@@ -386,7 +384,7 @@ int mlx4_en_create_rx_ring(struct mlx4_en_priv *priv, ...@@ -386,7 +384,7 @@ int mlx4_en_create_rx_ring(struct mlx4_en_priv *priv,
err = mlx4_en_map_buffer(&ring->wqres.buf); err = mlx4_en_map_buffer(&ring->wqres.buf);
if (err) { if (err) {
mlx4_err(mdev, "Failed to map RX buffer\n"); en_err(priv, "Failed to map RX buffer\n");
goto err_hwq; goto err_hwq;
} }
ring->buf = ring->wqres.buf.direct.buf; ring->buf = ring->wqres.buf.direct.buf;
...@@ -404,7 +402,7 @@ int mlx4_en_create_rx_ring(struct mlx4_en_priv *priv, ...@@ -404,7 +402,7 @@ int mlx4_en_create_rx_ring(struct mlx4_en_priv *priv,
sizeof(struct net_lro_desc), sizeof(struct net_lro_desc),
GFP_KERNEL); GFP_KERNEL);
if (!ring->lro.lro_arr) { if (!ring->lro.lro_arr) {
mlx4_err(mdev, "Failed to allocate lro array\n"); en_err(priv, "Failed to allocate lro array\n");
goto err_map; goto err_map;
} }
ring->lro.get_frag_header = mlx4_en_get_frag_header; ring->lro.get_frag_header = mlx4_en_get_frag_header;
...@@ -455,7 +453,7 @@ int mlx4_en_activate_rx_rings(struct mlx4_en_priv *priv) ...@@ -455,7 +453,7 @@ int mlx4_en_activate_rx_rings(struct mlx4_en_priv *priv)
/* Initialize page allocators */ /* Initialize page allocators */
err = mlx4_en_init_allocator(priv, ring); err = mlx4_en_init_allocator(priv, ring);
if (err) { if (err) {
mlx4_err(mdev, "Failed initializing ring allocator\n"); en_err(priv, "Failed initializing ring allocator\n");
ring_ind--; ring_ind--;
goto err_allocator; goto err_allocator;
} }
...@@ -486,7 +484,7 @@ int mlx4_en_activate_rx_rings(struct mlx4_en_priv *priv) ...@@ -486,7 +484,7 @@ int mlx4_en_activate_rx_rings(struct mlx4_en_priv *priv)
err = mlx4_srq_alloc(mdev->dev, mdev->priv_pdn, &ring->wqres.mtt, err = mlx4_srq_alloc(mdev->dev, mdev->priv_pdn, &ring->wqres.mtt,
ring->wqres.db.dma, &ring->srq); ring->wqres.db.dma, &ring->srq);
if (err){ if (err){
mlx4_err(mdev, "Failed to allocate srq\n"); en_err(priv, "Failed to allocate srq\n");
ring_ind--; ring_ind--;
goto err_srq; goto err_srq;
} }
...@@ -601,7 +599,7 @@ static struct sk_buff *mlx4_en_rx_skb(struct mlx4_en_priv *priv, ...@@ -601,7 +599,7 @@ static struct sk_buff *mlx4_en_rx_skb(struct mlx4_en_priv *priv,
skb = dev_alloc_skb(SMALL_PACKET_SIZE + NET_IP_ALIGN); skb = dev_alloc_skb(SMALL_PACKET_SIZE + NET_IP_ALIGN);
if (!skb) { if (!skb) {
mlx4_dbg(RX_ERR, priv, "Failed allocating skb\n"); en_dbg(RX_ERR, priv, "Failed allocating skb\n");
return NULL; return NULL;
} }
skb->dev = priv->dev; skb->dev = priv->dev;
...@@ -680,7 +678,6 @@ static void mlx4_en_copy_desc(struct mlx4_en_priv *priv, ...@@ -680,7 +678,6 @@ static void mlx4_en_copy_desc(struct mlx4_en_priv *priv,
int mlx4_en_process_rx_cq(struct net_device *dev, struct mlx4_en_cq *cq, int budget) int mlx4_en_process_rx_cq(struct net_device *dev, struct mlx4_en_cq *cq, int budget)
{ {
struct mlx4_en_priv *priv = netdev_priv(dev); struct mlx4_en_priv *priv = netdev_priv(dev);
struct mlx4_en_dev *mdev = priv->mdev;
struct mlx4_cqe *cqe; struct mlx4_cqe *cqe;
struct mlx4_en_rx_ring *ring = &priv->rx_ring[cq->ring]; struct mlx4_en_rx_ring *ring = &priv->rx_ring[cq->ring];
struct skb_frag_struct *skb_frags; struct skb_frag_struct *skb_frags;
...@@ -717,14 +714,14 @@ int mlx4_en_process_rx_cq(struct net_device *dev, struct mlx4_en_cq *cq, int bud ...@@ -717,14 +714,14 @@ int mlx4_en_process_rx_cq(struct net_device *dev, struct mlx4_en_cq *cq, int bud
/* Drop packet on bad receive or bad checksum */ /* Drop packet on bad receive or bad checksum */
if (unlikely((cqe->owner_sr_opcode & MLX4_CQE_OPCODE_MASK) == if (unlikely((cqe->owner_sr_opcode & MLX4_CQE_OPCODE_MASK) ==
MLX4_CQE_OPCODE_ERROR)) { MLX4_CQE_OPCODE_ERROR)) {
mlx4_err(mdev, "CQE completed in error - vendor " en_err(priv, "CQE completed in error - vendor "
"syndrom:%d syndrom:%d\n", "syndrom:%d syndrom:%d\n",
((struct mlx4_err_cqe *) cqe)->vendor_err_syndrome, ((struct mlx4_err_cqe *) cqe)->vendor_err_syndrome,
((struct mlx4_err_cqe *) cqe)->syndrome); ((struct mlx4_err_cqe *) cqe)->syndrome);
goto next; goto next;
} }
if (unlikely(cqe->badfcs_enc & MLX4_CQE_BAD_FCS)) { if (unlikely(cqe->badfcs_enc & MLX4_CQE_BAD_FCS)) {
mlx4_dbg(RX_ERR, priv, "Accepted frame with bad FCS\n"); en_dbg(RX_ERR, priv, "Accepted frame with bad FCS\n");
goto next; goto next;
} }
...@@ -874,7 +871,7 @@ static int mlx4_en_last_alloc_offset(struct mlx4_en_priv *priv, u16 stride, u16 ...@@ -874,7 +871,7 @@ static int mlx4_en_last_alloc_offset(struct mlx4_en_priv *priv, u16 stride, u16
u16 res = MLX4_EN_ALLOC_SIZE % stride; u16 res = MLX4_EN_ALLOC_SIZE % stride;
u16 offset = MLX4_EN_ALLOC_SIZE - stride - res + align; u16 offset = MLX4_EN_ALLOC_SIZE - stride - res + align;
mlx4_dbg(DRV, priv, "Calculated last offset for stride:%d align:%d " en_dbg(DRV, priv, "Calculated last offset for stride:%d align:%d "
"res:%d offset:%d\n", stride, align, res, offset); "res:%d offset:%d\n", stride, align, res, offset);
return offset; return offset;
} }
...@@ -919,10 +916,10 @@ void mlx4_en_calc_rx_buf(struct net_device *dev) ...@@ -919,10 +916,10 @@ void mlx4_en_calc_rx_buf(struct net_device *dev)
priv->rx_skb_size = eff_mtu; priv->rx_skb_size = eff_mtu;
priv->log_rx_info = ROUNDUP_LOG2(i * sizeof(struct skb_frag_struct)); priv->log_rx_info = ROUNDUP_LOG2(i * sizeof(struct skb_frag_struct));
mlx4_dbg(DRV, priv, "Rx buffer scatter-list (effective-mtu:%d " en_dbg(DRV, priv, "Rx buffer scatter-list (effective-mtu:%d "
"num_frags:%d):\n", eff_mtu, priv->num_frags); "num_frags:%d):\n", eff_mtu, priv->num_frags);
for (i = 0; i < priv->num_frags; i++) { for (i = 0; i < priv->num_frags; i++) {
mlx4_dbg(DRV, priv, " frag:%d - size:%d prefix:%d align:%d " en_dbg(DRV, priv, " frag:%d - size:%d prefix:%d align:%d "
"stride:%d last_offset:%d\n", i, "stride:%d last_offset:%d\n", i,
priv->frag_info[i].frag_size, priv->frag_info[i].frag_size,
priv->frag_info[i].frag_prefix_size, priv->frag_info[i].frag_prefix_size,
...@@ -942,12 +939,12 @@ void mlx4_en_set_default_rss_map(struct mlx4_en_priv *priv, ...@@ -942,12 +939,12 @@ void mlx4_en_set_default_rss_map(struct mlx4_en_priv *priv,
int i; int i;
rss_map->size = roundup_pow_of_two(num_entries); rss_map->size = roundup_pow_of_two(num_entries);
mlx4_dbg(DRV, priv, "Setting default RSS map of %d entires\n", en_dbg(DRV, priv, "Setting default RSS map of %d entires\n",
rss_map->size); rss_map->size);
for (i = 0; i < rss_map->size; i++) { for (i = 0; i < rss_map->size; i++) {
rss_map->map[i] = i % num_rings; rss_map->map[i] = i % num_rings;
mlx4_dbg(DRV, priv, "Entry %d ---> ring %d\n", i, rss_map->map[i]); en_dbg(DRV, priv, "Entry %d ---> ring %d\n", i, rss_map->map[i]);
} }
} }
...@@ -962,13 +959,13 @@ static int mlx4_en_config_rss_qp(struct mlx4_en_priv *priv, ...@@ -962,13 +959,13 @@ static int mlx4_en_config_rss_qp(struct mlx4_en_priv *priv,
context = kmalloc(sizeof *context , GFP_KERNEL); context = kmalloc(sizeof *context , GFP_KERNEL);
if (!context) { if (!context) {
mlx4_err(mdev, "Failed to allocate qp context\n"); en_err(priv, "Failed to allocate qp context\n");
return -ENOMEM; return -ENOMEM;
} }
err = mlx4_qp_alloc(mdev->dev, qpn, qp); err = mlx4_qp_alloc(mdev->dev, qpn, qp);
if (err) { if (err) {
mlx4_err(mdev, "Failed to allocate qp #%d\n", qpn); en_err(priv, "Failed to allocate qp #%x\n", qpn);
goto out; goto out;
} }
qp->event = mlx4_en_sqp_event; qp->event = mlx4_en_sqp_event;
...@@ -1000,12 +997,11 @@ int mlx4_en_config_rss_steer(struct mlx4_en_priv *priv) ...@@ -1000,12 +997,11 @@ int mlx4_en_config_rss_steer(struct mlx4_en_priv *priv)
int err = 0; int err = 0;
int good_qps = 0; int good_qps = 0;
mlx4_dbg(DRV, priv, "Configuring rss steering for port %u\n", priv->port); en_dbg(DRV, priv, "Configuring rss steering\n");
err = mlx4_qp_reserve_range(mdev->dev, rss_map->size, err = mlx4_qp_reserve_range(mdev->dev, rss_map->size,
rss_map->size, &rss_map->base_qpn); rss_map->size, &rss_map->base_qpn);
if (err) { if (err) {
mlx4_err(mdev, "Failed reserving %d qps for port %u\n", en_err(priv, "Failed reserving %d qps\n", rss_map->size);
rss_map->size, priv->port);
return err; return err;
} }
...@@ -1025,13 +1021,13 @@ int mlx4_en_config_rss_steer(struct mlx4_en_priv *priv) ...@@ -1025,13 +1021,13 @@ int mlx4_en_config_rss_steer(struct mlx4_en_priv *priv)
/* Configure RSS indirection qp */ /* Configure RSS indirection qp */
err = mlx4_qp_reserve_range(mdev->dev, 1, 1, &priv->base_qpn); err = mlx4_qp_reserve_range(mdev->dev, 1, 1, &priv->base_qpn);
if (err) { if (err) {
mlx4_err(mdev, "Failed to reserve range for RSS " en_err(priv, "Failed to reserve range for RSS "
"indirection qp\n"); "indirection qp\n");
goto rss_err; goto rss_err;
} }
err = mlx4_qp_alloc(mdev->dev, priv->base_qpn, &rss_map->indir_qp); err = mlx4_qp_alloc(mdev->dev, priv->base_qpn, &rss_map->indir_qp);
if (err) { if (err) {
mlx4_err(mdev, "Failed to allocate RSS indirection QP\n"); en_err(priv, "Failed to allocate RSS indirection QP\n");
goto reserve_err; goto reserve_err;
} }
rss_map->indir_qp.event = mlx4_en_sqp_event; rss_map->indir_qp.event = mlx4_en_sqp_event;
......
...@@ -68,15 +68,15 @@ int mlx4_en_create_tx_ring(struct mlx4_en_priv *priv, ...@@ -68,15 +68,15 @@ int mlx4_en_create_tx_ring(struct mlx4_en_priv *priv,
tmp = size * sizeof(struct mlx4_en_tx_info); tmp = size * sizeof(struct mlx4_en_tx_info);
ring->tx_info = vmalloc(tmp); ring->tx_info = vmalloc(tmp);
if (!ring->tx_info) { if (!ring->tx_info) {
mlx4_err(mdev, "Failed allocating tx_info ring\n"); en_err(priv, "Failed allocating tx_info ring\n");
return -ENOMEM; return -ENOMEM;
} }
mlx4_dbg(DRV, priv, "Allocated tx_info ring at addr:%p size:%d\n", en_dbg(DRV, priv, "Allocated tx_info ring at addr:%p size:%d\n",
ring->tx_info, tmp); ring->tx_info, tmp);
ring->bounce_buf = kmalloc(MAX_DESC_SIZE, GFP_KERNEL); ring->bounce_buf = kmalloc(MAX_DESC_SIZE, GFP_KERNEL);
if (!ring->bounce_buf) { if (!ring->bounce_buf) {
mlx4_err(mdev, "Failed allocating bounce buffer\n"); en_err(priv, "Failed allocating bounce buffer\n");
err = -ENOMEM; err = -ENOMEM;
goto err_tx; goto err_tx;
} }
...@@ -85,31 +85,31 @@ int mlx4_en_create_tx_ring(struct mlx4_en_priv *priv, ...@@ -85,31 +85,31 @@ int mlx4_en_create_tx_ring(struct mlx4_en_priv *priv,
err = mlx4_alloc_hwq_res(mdev->dev, &ring->wqres, ring->buf_size, err = mlx4_alloc_hwq_res(mdev->dev, &ring->wqres, ring->buf_size,
2 * PAGE_SIZE); 2 * PAGE_SIZE);
if (err) { if (err) {
mlx4_err(mdev, "Failed allocating hwq resources\n"); en_err(priv, "Failed allocating hwq resources\n");
goto err_bounce; goto err_bounce;
} }
err = mlx4_en_map_buffer(&ring->wqres.buf); err = mlx4_en_map_buffer(&ring->wqres.buf);
if (err) { if (err) {
mlx4_err(mdev, "Failed to map TX buffer\n"); en_err(priv, "Failed to map TX buffer\n");
goto err_hwq_res; goto err_hwq_res;
} }
ring->buf = ring->wqres.buf.direct.buf; ring->buf = ring->wqres.buf.direct.buf;
mlx4_dbg(DRV, priv, "Allocated TX ring (addr:%p) - buf:%p size:%d " en_dbg(DRV, priv, "Allocated TX ring (addr:%p) - buf:%p size:%d "
"buf_size:%d dma:%llx\n", ring, ring->buf, ring->size, "buf_size:%d dma:%llx\n", ring, ring->buf, ring->size,
ring->buf_size, (unsigned long long) ring->wqres.buf.direct.map); ring->buf_size, (unsigned long long) ring->wqres.buf.direct.map);
err = mlx4_qp_reserve_range(mdev->dev, 1, 1, &ring->qpn); err = mlx4_qp_reserve_range(mdev->dev, 1, 1, &ring->qpn);
if (err) { if (err) {
mlx4_err(mdev, "Failed reserving qp for tx ring.\n"); en_err(priv, "Failed reserving qp for tx ring.\n");
goto err_map; goto err_map;
} }
err = mlx4_qp_alloc(mdev->dev, ring->qpn, &ring->qp); err = mlx4_qp_alloc(mdev->dev, ring->qpn, &ring->qp);
if (err) { if (err) {
mlx4_err(mdev, "Failed allocating qp %d\n", ring->qpn); en_err(priv, "Failed allocating qp %d\n", ring->qpn);
goto err_reserve; goto err_reserve;
} }
ring->qp.event = mlx4_en_sqp_event; ring->qp.event = mlx4_en_sqp_event;
...@@ -135,7 +135,7 @@ void mlx4_en_destroy_tx_ring(struct mlx4_en_priv *priv, ...@@ -135,7 +135,7 @@ void mlx4_en_destroy_tx_ring(struct mlx4_en_priv *priv,
struct mlx4_en_tx_ring *ring) struct mlx4_en_tx_ring *ring)
{ {
struct mlx4_en_dev *mdev = priv->mdev; struct mlx4_en_dev *mdev = priv->mdev;
mlx4_dbg(DRV, priv, "Destroying tx ring, qpn: %d\n", ring->qpn); en_dbg(DRV, priv, "Destroying tx ring, qpn: %d\n", ring->qpn);
mlx4_qp_remove(mdev->dev, &ring->qp); mlx4_qp_remove(mdev->dev, &ring->qp);
mlx4_qp_free(mdev->dev, &ring->qp); mlx4_qp_free(mdev->dev, &ring->qp);
...@@ -274,12 +274,12 @@ int mlx4_en_free_tx_buf(struct net_device *dev, struct mlx4_en_tx_ring *ring) ...@@ -274,12 +274,12 @@ int mlx4_en_free_tx_buf(struct net_device *dev, struct mlx4_en_tx_ring *ring)
/* Skip last polled descriptor */ /* Skip last polled descriptor */
ring->cons += ring->last_nr_txbb; ring->cons += ring->last_nr_txbb;
mlx4_dbg(DRV, priv, "Freeing Tx buf - cons:0x%x prod:0x%x\n", en_dbg(DRV, priv, "Freeing Tx buf - cons:0x%x prod:0x%x\n",
ring->cons, ring->prod); ring->cons, ring->prod);
if ((u32) (ring->prod - ring->cons) > ring->size) { if ((u32) (ring->prod - ring->cons) > ring->size) {
if (netif_msg_tx_err(priv)) if (netif_msg_tx_err(priv))
mlx4_warn(priv->mdev, "Tx consumer passed producer!\n"); en_warn(priv, "Tx consumer passed producer!\n");
return 0; return 0;
} }
...@@ -292,7 +292,7 @@ int mlx4_en_free_tx_buf(struct net_device *dev, struct mlx4_en_tx_ring *ring) ...@@ -292,7 +292,7 @@ int mlx4_en_free_tx_buf(struct net_device *dev, struct mlx4_en_tx_ring *ring)
} }
if (cnt) if (cnt)
mlx4_dbg(DRV, priv, "Freed %d uncompleted tx descriptors\n", cnt); en_dbg(DRV, priv, "Freed %d uncompleted tx descriptors\n", cnt);
return cnt; return cnt;
} }
...@@ -321,7 +321,7 @@ void mlx4_en_set_prio_map(struct mlx4_en_priv *priv, u16 *prio_map, u32 ring_num ...@@ -321,7 +321,7 @@ void mlx4_en_set_prio_map(struct mlx4_en_priv *priv, u16 *prio_map, u32 ring_num
num = 0; num = 0;
} }
prio_map[prio] = ring; prio_map[prio] = ring;
mlx4_dbg(DRV, priv, " prio:%d --> ring:%d\n", prio, ring); en_dbg(DRV, priv, " prio:%d --> ring:%d\n", prio, ring);
num++; num++;
} }
} }
...@@ -539,7 +539,6 @@ static int get_real_size(struct sk_buff *skb, struct net_device *dev, ...@@ -539,7 +539,6 @@ static int get_real_size(struct sk_buff *skb, struct net_device *dev,
int *lso_header_size) int *lso_header_size)
{ {
struct mlx4_en_priv *priv = netdev_priv(dev); struct mlx4_en_priv *priv = netdev_priv(dev);
struct mlx4_en_dev *mdev = priv->mdev;
int real_size; int real_size;
if (skb_is_gso(skb)) { if (skb_is_gso(skb)) {
...@@ -553,14 +552,14 @@ static int get_real_size(struct sk_buff *skb, struct net_device *dev, ...@@ -553,14 +552,14 @@ static int get_real_size(struct sk_buff *skb, struct net_device *dev,
real_size += DS_SIZE; real_size += DS_SIZE;
else { else {
if (netif_msg_tx_err(priv)) if (netif_msg_tx_err(priv))
mlx4_warn(mdev, "Non-linear headers\n"); en_warn(priv, "Non-linear headers\n");
dev_kfree_skb_any(skb); dev_kfree_skb_any(skb);
return 0; return 0;
} }
} }
if (unlikely(*lso_header_size > MAX_LSO_HDR_SIZE)) { if (unlikely(*lso_header_size > MAX_LSO_HDR_SIZE)) {
if (netif_msg_tx_err(priv)) if (netif_msg_tx_err(priv))
mlx4_warn(mdev, "LSO header size too big\n"); en_warn(priv, "LSO header size too big\n");
dev_kfree_skb_any(skb); dev_kfree_skb_any(skb);
return 0; return 0;
} }
...@@ -669,7 +668,7 @@ int mlx4_en_xmit(struct sk_buff *skb, struct net_device *dev) ...@@ -669,7 +668,7 @@ int mlx4_en_xmit(struct sk_buff *skb, struct net_device *dev)
nr_txbb = desc_size / TXBB_SIZE; nr_txbb = desc_size / TXBB_SIZE;
if (unlikely(nr_txbb > MAX_DESC_TXBBS)) { if (unlikely(nr_txbb > MAX_DESC_TXBBS)) {
if (netif_msg_tx_err(priv)) if (netif_msg_tx_err(priv))
mlx4_warn(mdev, "Oversized header or SG list\n"); en_warn(priv, "Oversized header or SG list\n");
dev_kfree_skb_any(skb); dev_kfree_skb_any(skb);
return NETDEV_TX_OK; return NETDEV_TX_OK;
} }
...@@ -695,7 +694,7 @@ int mlx4_en_xmit(struct sk_buff *skb, struct net_device *dev) ...@@ -695,7 +694,7 @@ int mlx4_en_xmit(struct sk_buff *skb, struct net_device *dev)
/* Now that we know what Tx ring to use */ /* Now that we know what Tx ring to use */
if (unlikely(!priv->port_up)) { if (unlikely(!priv->port_up)) {
if (netif_msg_tx_err(priv)) if (netif_msg_tx_err(priv))
mlx4_warn(mdev, "xmit: port down!\n"); en_warn(priv, "xmit: port down!\n");
dev_kfree_skb_any(skb); dev_kfree_skb_any(skb);
return NETDEV_TX_OK; return NETDEV_TX_OK;
} }
......
...@@ -55,20 +55,36 @@ ...@@ -55,20 +55,36 @@
#define MLX4_EN_MSG_LEVEL (NETIF_MSG_LINK | NETIF_MSG_IFDOWN) #define MLX4_EN_MSG_LEVEL (NETIF_MSG_LINK | NETIF_MSG_IFDOWN)
#define mlx4_dbg(mlevel, priv, format, arg...) \ #define en_print(level, priv, format, arg...) \
{ \
if ((priv)->registered) \
printk(level "%s: %s: " format, DRV_NAME, \
(priv->dev)->name, ## arg); \
else \
printk(level "%s: %s: Port %d: " format, \
DRV_NAME, dev_name(&priv->mdev->pdev->dev), \
(priv)->port, ## arg); \
}
#define en_dbg(mlevel, priv, format, arg...) \
{ \
if (NETIF_MSG_##mlevel & priv->msg_enable) \ if (NETIF_MSG_##mlevel & priv->msg_enable) \
printk(KERN_DEBUG "%s %s: " format , DRV_NAME ,\ en_print(KERN_DEBUG, priv, format, ## arg) \
(dev_name(&priv->mdev->pdev->dev)) , ## arg) }
#define en_warn(priv, format, arg...) \
en_print(KERN_WARNING, priv, format, ## arg)
#define en_err(priv, format, arg...) \
en_print(KERN_ERR, priv, format, ## arg)
#define mlx4_err(mdev, format, arg...) \ #define mlx4_err(mdev, format, arg...) \
printk(KERN_ERR "%s %s: " format , DRV_NAME ,\ printk(KERN_ERR "%s %s: " format , DRV_NAME ,\
(dev_name(&mdev->pdev->dev)) , ## arg) dev_name(&mdev->pdev->dev) , ## arg)
#define mlx4_info(mdev, format, arg...) \ #define mlx4_info(mdev, format, arg...) \
printk(KERN_INFO "%s %s: " format , DRV_NAME ,\ printk(KERN_INFO "%s %s: " format , DRV_NAME ,\
(dev_name(&mdev->pdev->dev)) , ## arg) dev_name(&mdev->pdev->dev) , ## arg)
#define mlx4_warn(mdev, format, arg...) \ #define mlx4_warn(mdev, format, arg...) \
printk(KERN_WARNING "%s %s: " format , DRV_NAME ,\ printk(KERN_WARNING "%s %s: " format , DRV_NAME ,\
(dev_name(&mdev->pdev->dev)) , ## arg) dev_name(&mdev->pdev->dev) , ## arg)
/* /*
* Device constants * Device constants
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment