Commit 21c2dece authored by Santiago Leon's avatar Santiago Leon Committed by David S. Miller

ibmveth: Convert driver specific error functions to netdev_err

Use netdev_err to standardise the error output.
Signed-off-by: default avatarAnton Blanchard <anton@samba.org>
Signed-off-by: default avatarSantiago Leon <santil@linux.vnet.ibm.com>
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parent c43ced18
...@@ -57,12 +57,6 @@ ...@@ -57,12 +57,6 @@
#undef DEBUG #undef DEBUG
#define ibmveth_printk(fmt, args...) \
printk(KERN_DEBUG "%s: " fmt, __FILE__, ## args)
#define ibmveth_error_printk(fmt, args...) \
printk(KERN_ERR "(%s:%3.3d ua:%x) ERROR: " fmt, __FILE__, __LINE__ , adapter->vdev->unit_address, ## args)
#ifdef DEBUG #ifdef DEBUG
#define ibmveth_assert(expr) \ #define ibmveth_assert(expr) \
if(!(expr)) { \ if(!(expr)) { \
...@@ -558,7 +552,8 @@ static int ibmveth_open(struct net_device *netdev) ...@@ -558,7 +552,8 @@ static int ibmveth_open(struct net_device *netdev)
adapter->filter_list_addr = (void*) get_zeroed_page(GFP_KERNEL); adapter->filter_list_addr = (void*) get_zeroed_page(GFP_KERNEL);
if(!adapter->buffer_list_addr || !adapter->filter_list_addr) { if(!adapter->buffer_list_addr || !adapter->filter_list_addr) {
ibmveth_error_printk("unable to allocate filter or buffer list pages\n"); netdev_err(netdev, "unable to allocate filter or buffer list "
"pages\n");
ibmveth_cleanup(adapter); ibmveth_cleanup(adapter);
napi_disable(&adapter->napi); napi_disable(&adapter->napi);
return -ENOMEM; return -ENOMEM;
...@@ -568,7 +563,7 @@ static int ibmveth_open(struct net_device *netdev) ...@@ -568,7 +563,7 @@ static int ibmveth_open(struct net_device *netdev)
adapter->rx_queue.queue_addr = kmalloc(adapter->rx_queue.queue_len, GFP_KERNEL); adapter->rx_queue.queue_addr = kmalloc(adapter->rx_queue.queue_len, GFP_KERNEL);
if(!adapter->rx_queue.queue_addr) { if(!adapter->rx_queue.queue_addr) {
ibmveth_error_printk("unable to allocate rx queue pages\n"); netdev_err(netdev, "unable to allocate rx queue pages\n");
ibmveth_cleanup(adapter); ibmveth_cleanup(adapter);
napi_disable(&adapter->napi); napi_disable(&adapter->napi);
return -ENOMEM; return -ENOMEM;
...@@ -587,7 +582,8 @@ static int ibmveth_open(struct net_device *netdev) ...@@ -587,7 +582,8 @@ static int ibmveth_open(struct net_device *netdev)
if ((dma_mapping_error(dev, adapter->buffer_list_dma)) || if ((dma_mapping_error(dev, adapter->buffer_list_dma)) ||
(dma_mapping_error(dev, adapter->filter_list_dma)) || (dma_mapping_error(dev, adapter->filter_list_dma)) ||
(dma_mapping_error(dev, adapter->rx_queue.queue_dma))) { (dma_mapping_error(dev, adapter->rx_queue.queue_dma))) {
ibmveth_error_printk("unable to map filter or buffer list pages\n"); netdev_err(netdev, "unable to map filter or buffer list "
"pages\n");
ibmveth_cleanup(adapter); ibmveth_cleanup(adapter);
napi_disable(&adapter->napi); napi_disable(&adapter->napi);
return -ENOMEM; return -ENOMEM;
...@@ -612,8 +608,10 @@ static int ibmveth_open(struct net_device *netdev) ...@@ -612,8 +608,10 @@ static int ibmveth_open(struct net_device *netdev)
lpar_rc = ibmveth_register_logical_lan(adapter, rxq_desc, mac_address); lpar_rc = ibmveth_register_logical_lan(adapter, rxq_desc, mac_address);
if(lpar_rc != H_SUCCESS) { if(lpar_rc != H_SUCCESS) {
ibmveth_error_printk("h_register_logical_lan failed with %ld\n", lpar_rc); netdev_err(netdev, "h_register_logical_lan failed with %ld\n",
ibmveth_error_printk("buffer TCE:0x%llx filter TCE:0x%llx rxq desc:0x%llx MAC:0x%llx\n", lpar_rc);
netdev_err(netdev, "buffer TCE:0x%llx filter TCE:0x%llx rxq "
"desc:0x%llx MAC:0x%llx\n",
adapter->buffer_list_dma, adapter->buffer_list_dma,
adapter->filter_list_dma, adapter->filter_list_dma,
rxq_desc.desc, rxq_desc.desc,
...@@ -627,7 +625,7 @@ static int ibmveth_open(struct net_device *netdev) ...@@ -627,7 +625,7 @@ static int ibmveth_open(struct net_device *netdev)
if(!adapter->rx_buff_pool[i].active) if(!adapter->rx_buff_pool[i].active)
continue; continue;
if (ibmveth_alloc_buffer_pool(&adapter->rx_buff_pool[i])) { if (ibmveth_alloc_buffer_pool(&adapter->rx_buff_pool[i])) {
ibmveth_error_printk("unable to alloc pool\n"); netdev_err(netdev, "unable to alloc pool\n");
adapter->rx_buff_pool[i].active = 0; adapter->rx_buff_pool[i].active = 0;
ibmveth_cleanup(adapter); ibmveth_cleanup(adapter);
napi_disable(&adapter->napi); napi_disable(&adapter->napi);
...@@ -637,7 +635,8 @@ static int ibmveth_open(struct net_device *netdev) ...@@ -637,7 +635,8 @@ static int ibmveth_open(struct net_device *netdev)
netdev_dbg(netdev, "registering irq 0x%x\n", netdev->irq); netdev_dbg(netdev, "registering irq 0x%x\n", netdev->irq);
if((rc = request_irq(netdev->irq, ibmveth_interrupt, 0, netdev->name, netdev)) != 0) { if((rc = request_irq(netdev->irq, ibmveth_interrupt, 0, netdev->name, netdev)) != 0) {
ibmveth_error_printk("unable to request irq 0x%x, rc %d\n", netdev->irq, rc); netdev_err(netdev, "unable to request irq 0x%x, rc %d\n",
netdev->irq, rc);
do { do {
rc = h_free_logical_lan(adapter->vdev->unit_address); rc = h_free_logical_lan(adapter->vdev->unit_address);
} while (H_IS_LONG_BUSY(rc) || (rc == H_BUSY)); } while (H_IS_LONG_BUSY(rc) || (rc == H_BUSY));
...@@ -650,7 +649,7 @@ static int ibmveth_open(struct net_device *netdev) ...@@ -650,7 +649,7 @@ static int ibmveth_open(struct net_device *netdev)
adapter->bounce_buffer = adapter->bounce_buffer =
kmalloc(netdev->mtu + IBMVETH_BUFF_OH, GFP_KERNEL); kmalloc(netdev->mtu + IBMVETH_BUFF_OH, GFP_KERNEL);
if (!adapter->bounce_buffer) { if (!adapter->bounce_buffer) {
ibmveth_error_printk("unable to allocate bounce buffer\n"); netdev_err(netdev, "unable to allocate bounce buffer\n");
ibmveth_cleanup(adapter); ibmveth_cleanup(adapter);
napi_disable(&adapter->napi); napi_disable(&adapter->napi);
return -ENOMEM; return -ENOMEM;
...@@ -659,7 +658,7 @@ static int ibmveth_open(struct net_device *netdev) ...@@ -659,7 +658,7 @@ static int ibmveth_open(struct net_device *netdev)
dma_map_single(&adapter->vdev->dev, adapter->bounce_buffer, dma_map_single(&adapter->vdev->dev, adapter->bounce_buffer,
netdev->mtu + IBMVETH_BUFF_OH, DMA_BIDIRECTIONAL); netdev->mtu + IBMVETH_BUFF_OH, DMA_BIDIRECTIONAL);
if (dma_mapping_error(dev, adapter->bounce_buffer_dma)) { if (dma_mapping_error(dev, adapter->bounce_buffer_dma)) {
ibmveth_error_printk("unable to map bounce buffer\n"); netdev_err(netdev, "unable to map bounce buffer\n");
ibmveth_cleanup(adapter); ibmveth_cleanup(adapter);
napi_disable(&adapter->napi); napi_disable(&adapter->napi);
return -ENOMEM; return -ENOMEM;
...@@ -695,8 +694,8 @@ static int ibmveth_close(struct net_device *netdev) ...@@ -695,8 +694,8 @@ static int ibmveth_close(struct net_device *netdev)
if(lpar_rc != H_SUCCESS) if(lpar_rc != H_SUCCESS)
{ {
ibmveth_error_printk("h_free_logical_lan failed with %lx, continuing with close\n", netdev_err(netdev, "h_free_logical_lan failed with %lx, "
lpar_rc); "continuing with close\n", lpar_rc);
} }
free_irq(netdev->irq, netdev); free_irq(netdev->irq, netdev);
...@@ -806,7 +805,7 @@ static int ibmveth_set_csum_offload(struct net_device *dev, u32 data, ...@@ -806,7 +805,7 @@ static int ibmveth_set_csum_offload(struct net_device *dev, u32 data,
set_attr, &ret_attr); set_attr, &ret_attr);
if (ret != H_SUCCESS) { if (ret != H_SUCCESS) {
ibmveth_error_printk("unable to change IPv4 checksum " netdev_err(dev, "unable to change IPv4 checksum "
"offload settings. %d rc=%ld\n", "offload settings. %d rc=%ld\n",
data, ret); data, ret);
...@@ -819,7 +818,7 @@ static int ibmveth_set_csum_offload(struct net_device *dev, u32 data, ...@@ -819,7 +818,7 @@ static int ibmveth_set_csum_offload(struct net_device *dev, u32 data,
clr_attr6, set_attr6, &ret_attr); clr_attr6, set_attr6, &ret_attr);
if (ret6 != H_SUCCESS) { if (ret6 != H_SUCCESS) {
ibmveth_error_printk("unable to change IPv6 checksum " netdev_err(dev, "unable to change IPv6 checksum "
"offload settings. %d rc=%ld\n", "offload settings. %d rc=%ld\n",
data, ret); data, ret);
...@@ -835,8 +834,9 @@ static int ibmveth_set_csum_offload(struct net_device *dev, u32 data, ...@@ -835,8 +834,9 @@ static int ibmveth_set_csum_offload(struct net_device *dev, u32 data,
rc1 = -EIO; rc1 = -EIO;
} else { } else {
rc1 = -EIO; rc1 = -EIO;
ibmveth_error_printk("unable to change checksum offload settings." netdev_err(dev, "unable to change checksum offload settings."
" %d rc=%ld ret_attr=%lx\n", data, ret, ret_attr); " %d rc=%ld ret_attr=%lx\n", data, ret,
ret_attr);
} }
if (restart) if (restart)
...@@ -952,8 +952,8 @@ static int ibmveth_send(struct ibmveth_adapter *adapter, ...@@ -952,8 +952,8 @@ static int ibmveth_send(struct ibmveth_adapter *adapter,
} while ((ret == H_BUSY) && (retry_count--)); } while ((ret == H_BUSY) && (retry_count--));
if (ret != H_SUCCESS && ret != H_DROPPED) { if (ret != H_SUCCESS && ret != H_DROPPED) {
ibmveth_error_printk("tx: h_send_logical_lan failed with " netdev_err(adapter->netdev, "tx: h_send_logical_lan failed "
"rc=%ld\n", ret); "with rc=%ld\n", ret);
return 1; return 1;
} }
...@@ -986,7 +986,7 @@ static netdev_tx_t ibmveth_start_xmit(struct sk_buff *skb, ...@@ -986,7 +986,7 @@ static netdev_tx_t ibmveth_start_xmit(struct sk_buff *skb,
ipv6_hdr(skb)->nexthdr != IPPROTO_TCP)) && ipv6_hdr(skb)->nexthdr != IPPROTO_TCP)) &&
skb_checksum_help(skb)) { skb_checksum_help(skb)) {
ibmveth_error_printk("tx: failed to checksum packet\n"); netdev_err(netdev, "tx: failed to checksum packet\n");
netdev->stats.tx_dropped++; netdev->stats.tx_dropped++;
goto out; goto out;
} }
...@@ -1082,7 +1082,7 @@ static netdev_tx_t ibmveth_start_xmit(struct sk_buff *skb, ...@@ -1082,7 +1082,7 @@ static netdev_tx_t ibmveth_start_xmit(struct sk_buff *skb,
map_failed: map_failed:
if (!firmware_has_feature(FW_FEATURE_CMO)) if (!firmware_has_feature(FW_FEATURE_CMO))
ibmveth_error_printk("tx: unable to map xmit buffer\n"); netdev_err(netdev, "tx: unable to map xmit buffer\n");
adapter->tx_map_failed++; adapter->tx_map_failed++;
skb_linearize(skb); skb_linearize(skb);
force_bounce = 1; force_bounce = 1;
...@@ -1198,7 +1198,8 @@ static void ibmveth_set_multicast_list(struct net_device *netdev) ...@@ -1198,7 +1198,8 @@ static void ibmveth_set_multicast_list(struct net_device *netdev)
IbmVethMcastDisableFiltering, IbmVethMcastDisableFiltering,
0); 0);
if(lpar_rc != H_SUCCESS) { if(lpar_rc != H_SUCCESS) {
ibmveth_error_printk("h_multicast_ctrl rc=%ld when entering promisc mode\n", lpar_rc); netdev_err(netdev, "h_multicast_ctrl rc=%ld when "
"entering promisc mode\n", lpar_rc);
} }
} else { } else {
struct netdev_hw_addr *ha; struct netdev_hw_addr *ha;
...@@ -1209,7 +1210,9 @@ static void ibmveth_set_multicast_list(struct net_device *netdev) ...@@ -1209,7 +1210,9 @@ static void ibmveth_set_multicast_list(struct net_device *netdev)
IbmVethMcastClearFilterTable, IbmVethMcastClearFilterTable,
0); 0);
if(lpar_rc != H_SUCCESS) { if(lpar_rc != H_SUCCESS) {
ibmveth_error_printk("h_multicast_ctrl rc=%ld when attempting to clear filter table\n", lpar_rc); netdev_err(netdev, "h_multicast_ctrl rc=%ld when "
"attempting to clear filter table\n",
lpar_rc);
} }
/* add the addresses to the filter table */ /* add the addresses to the filter table */
netdev_for_each_mc_addr(ha, netdev) { netdev_for_each_mc_addr(ha, netdev) {
...@@ -1220,7 +1223,9 @@ static void ibmveth_set_multicast_list(struct net_device *netdev) ...@@ -1220,7 +1223,9 @@ static void ibmveth_set_multicast_list(struct net_device *netdev)
IbmVethMcastAddFilter, IbmVethMcastAddFilter,
mcast_addr); mcast_addr);
if(lpar_rc != H_SUCCESS) { if(lpar_rc != H_SUCCESS) {
ibmveth_error_printk("h_multicast_ctrl rc=%ld when adding an entry to the filter table\n", lpar_rc); netdev_err(netdev, "h_multicast_ctrl rc=%ld "
"when adding an entry to the filter "
"table\n", lpar_rc);
} }
} }
...@@ -1229,7 +1234,8 @@ static void ibmveth_set_multicast_list(struct net_device *netdev) ...@@ -1229,7 +1234,8 @@ static void ibmveth_set_multicast_list(struct net_device *netdev)
IbmVethMcastEnableFiltering, IbmVethMcastEnableFiltering,
0); 0);
if(lpar_rc != H_SUCCESS) { if(lpar_rc != H_SUCCESS) {
ibmveth_error_printk("h_multicast_ctrl rc=%ld when enabling filtering\n", lpar_rc); netdev_err(netdev, "h_multicast_ctrl rc=%ld when "
"enabling filtering\n", lpar_rc);
} }
} }
} }
...@@ -1361,17 +1367,15 @@ static int __devinit ibmveth_probe(struct vio_dev *dev, const struct vio_device_ ...@@ -1361,17 +1367,15 @@ static int __devinit ibmveth_probe(struct vio_dev *dev, const struct vio_device_
mac_addr_p = (unsigned char *) vio_get_attribute(dev, mac_addr_p = (unsigned char *) vio_get_attribute(dev,
VETH_MAC_ADDR, NULL); VETH_MAC_ADDR, NULL);
if(!mac_addr_p) { if(!mac_addr_p) {
printk(KERN_ERR "(%s:%3.3d) ERROR: Can't find VETH_MAC_ADDR " dev_err(&dev->dev, "Can't find VETH_MAC_ADDR attribute\n");
"attribute\n", __FILE__, __LINE__);
return 0; return 0;
} }
mcastFilterSize_p = (unsigned int *) vio_get_attribute(dev, mcastFilterSize_p = (unsigned int *) vio_get_attribute(dev,
VETH_MCAST_FILTER_SIZE, NULL); VETH_MCAST_FILTER_SIZE, NULL);
if(!mcastFilterSize_p) { if(!mcastFilterSize_p) {
printk(KERN_ERR "(%s:%3.3d) ERROR: Can't find " dev_err(&dev->dev, "Can't find VETH_MCAST_FILTER_SIZE "
"VETH_MCAST_FILTER_SIZE attribute\n", "attribute\n");
__FILE__, __LINE__);
return 0; return 0;
} }
...@@ -1501,7 +1505,8 @@ const char * buf, size_t count) ...@@ -1501,7 +1505,8 @@ const char * buf, size_t count)
if (value && !pool->active) { if (value && !pool->active) {
if (netif_running(netdev)) { if (netif_running(netdev)) {
if(ibmveth_alloc_buffer_pool(pool)) { if(ibmveth_alloc_buffer_pool(pool)) {
ibmveth_error_printk("unable to alloc pool\n"); netdev_err(netdev,
"unable to alloc pool\n");
return -ENOMEM; return -ENOMEM;
} }
pool->active = 1; pool->active = 1;
...@@ -1527,7 +1532,7 @@ const char * buf, size_t count) ...@@ -1527,7 +1532,7 @@ const char * buf, size_t count)
} }
if (i == IbmVethNumBufferPools) { if (i == IbmVethNumBufferPools) {
ibmveth_error_printk("no active pool >= MTU\n"); netdev_err(netdev, "no active pool >= MTU\n");
return -EPERM; return -EPERM;
} }
...@@ -1635,7 +1640,8 @@ static struct vio_driver ibmveth_driver = { ...@@ -1635,7 +1640,8 @@ static struct vio_driver ibmveth_driver = {
static int __init ibmveth_module_init(void) static int __init ibmveth_module_init(void)
{ {
ibmveth_printk("%s: %s %s\n", ibmveth_driver_name, ibmveth_driver_string, ibmveth_driver_version); printk(KERN_DEBUG "%s: %s %s\n", ibmveth_driver_name,
ibmveth_driver_string, ibmveth_driver_version);
return vio_register_driver(&ibmveth_driver); return vio_register_driver(&ibmveth_driver);
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment