| /** |
| * This file is part of the Chelsio T4/T5/T6 Ethernet driver for Linux. |
| * |
| * Copyright (C) 2011-2016 Chelsio Communications. All rights reserved. |
| * |
| * This program is free software; you can redistribute it and/or modify |
| * it under the terms of the GNU General Public License as published by |
| * the Free Software Foundation. |
| * |
| * Written and Maintained by: |
| * Manoj Malviya (manojmalviya@chelsio.com) |
| * Atul Gupta (atul.gupta@chelsio.com) |
| * Jitendra Lulla (jlulla@chelsio.com) |
| * Yeshaswi M R Gowda (yeshaswi@chelsio.com) |
| * Harsh Jain (harsh@chelsio.com) |
| */ |
| |
| #include <linux/kernel.h> |
| #include <linux/module.h> |
| #include <linux/skbuff.h> |
| |
| #include <crypto/aes.h> |
| #include <crypto/hash.h> |
| |
| #include "t4_msg.h" |
| #include "chcr_core.h" |
| #include "cxgb4_uld.h" |
| |
| static LIST_HEAD(uld_ctx_list); |
| static DEFINE_MUTEX(dev_mutex); |
| static atomic_t dev_count; |
| static struct uld_ctx *ctx_rr; |
| |
| typedef int (*chcr_handler_func)(struct chcr_dev *dev, unsigned char *input); |
| static int cpl_fw6_pld_handler(struct chcr_dev *dev, unsigned char *input); |
| static void *chcr_uld_add(const struct cxgb4_lld_info *lld); |
| static int chcr_uld_state_change(void *handle, enum cxgb4_state state); |
| |
| static chcr_handler_func work_handlers[NUM_CPL_CMDS] = { |
| [CPL_FW6_PLD] = cpl_fw6_pld_handler, |
| }; |
| |
| static struct cxgb4_uld_info chcr_uld_info = { |
| .name = DRV_MODULE_NAME, |
| .nrxq = MAX_ULD_QSETS, |
| .ntxq = MAX_ULD_QSETS, |
| .rxq_size = 1024, |
| .add = chcr_uld_add, |
| .state_change = chcr_uld_state_change, |
| .rx_handler = chcr_uld_rx_handler, |
| #ifdef CONFIG_CHELSIO_IPSEC_INLINE |
| .tx_handler = chcr_uld_tx_handler, |
| #endif /* CONFIG_CHELSIO_IPSEC_INLINE */ |
| }; |
| |
| struct uld_ctx *assign_chcr_device(void) |
| { |
| struct uld_ctx *u_ctx = NULL; |
| |
| /* |
| * When multiple devices are present in system select |
| * device in round-robin fashion for crypto operations |
| * Although One session must use the same device to |
| * maintain request-response ordering. |
| */ |
| mutex_lock(&dev_mutex); |
| if (!list_empty(&uld_ctx_list)) { |
| u_ctx = ctx_rr; |
| if (list_is_last(&ctx_rr->entry, &uld_ctx_list)) |
| ctx_rr = list_first_entry(&uld_ctx_list, |
| struct uld_ctx, |
| entry); |
| else |
| ctx_rr = list_next_entry(ctx_rr, entry); |
| } |
| mutex_unlock(&dev_mutex); |
| return u_ctx; |
| } |
| |
| static int chcr_dev_add(struct uld_ctx *u_ctx) |
| { |
| struct chcr_dev *dev; |
| |
| dev = kzalloc(sizeof(*dev), GFP_KERNEL); |
| if (!dev) |
| return -ENXIO; |
| |
| spin_lock_init(&dev->lock_chcr_dev); |
| u_ctx->dev = dev; |
| dev->u_ctx = u_ctx; |
| atomic_inc(&dev_count); |
| mutex_lock(&dev_mutex); |
| list_add_tail(&u_ctx->entry, &uld_ctx_list); |
| if (!ctx_rr) |
| ctx_rr = u_ctx; |
| mutex_unlock(&dev_mutex); |
| return 0; |
| } |
| |
| static int chcr_dev_remove(struct uld_ctx *u_ctx) |
| { |
| if (ctx_rr == u_ctx) { |
| if (list_is_last(&ctx_rr->entry, &uld_ctx_list)) |
| ctx_rr = list_first_entry(&uld_ctx_list, |
| struct uld_ctx, |
| entry); |
| else |
| ctx_rr = list_next_entry(ctx_rr, entry); |
| } |
| list_del(&u_ctx->entry); |
| if (list_empty(&uld_ctx_list)) |
| ctx_rr = NULL; |
| kfree(u_ctx->dev); |
| u_ctx->dev = NULL; |
| atomic_dec(&dev_count); |
| return 0; |
| } |
| |
| static int cpl_fw6_pld_handler(struct chcr_dev *dev, |
| unsigned char *input) |
| { |
| struct crypto_async_request *req; |
| struct cpl_fw6_pld *fw6_pld; |
| u32 ack_err_status = 0; |
| int error_status = 0; |
| struct adapter *adap = padap(dev); |
| |
| fw6_pld = (struct cpl_fw6_pld *)input; |
| req = (struct crypto_async_request *)(uintptr_t)be64_to_cpu( |
| fw6_pld->data[1]); |
| |
| ack_err_status = |
| ntohl(*(__be32 *)((unsigned char *)&fw6_pld->data[0] + 4)); |
| if (ack_err_status) { |
| if (CHK_MAC_ERR_BIT(ack_err_status) || |
| CHK_PAD_ERR_BIT(ack_err_status)) |
| error_status = -EBADMSG; |
| atomic_inc(&adap->chcr_stats.error); |
| } |
| /* call completion callback with failure status */ |
| if (req) { |
| error_status = chcr_handle_resp(req, input, error_status); |
| } else { |
| pr_err("Incorrect request address from the firmware\n"); |
| return -EFAULT; |
| } |
| return 0; |
| } |
| |
| int chcr_send_wr(struct sk_buff *skb) |
| { |
| return cxgb4_crypto_send(skb->dev, skb); |
| } |
| |
| static void *chcr_uld_add(const struct cxgb4_lld_info *lld) |
| { |
| struct uld_ctx *u_ctx; |
| |
| /* Create the device and add it in the device list */ |
| if (!(lld->ulp_crypto & ULP_CRYPTO_LOOKASIDE)) |
| return ERR_PTR(-EOPNOTSUPP); |
| |
| /* Create the device and add it in the device list */ |
| u_ctx = kzalloc(sizeof(*u_ctx), GFP_KERNEL); |
| if (!u_ctx) { |
| u_ctx = ERR_PTR(-ENOMEM); |
| goto out; |
| } |
| u_ctx->lldi = *lld; |
| #ifdef CONFIG_CHELSIO_IPSEC_INLINE |
| if (lld->crypto & ULP_CRYPTO_IPSEC_INLINE) |
| chcr_add_xfrmops(lld); |
| #endif /* CONFIG_CHELSIO_IPSEC_INLINE */ |
| out: |
| return u_ctx; |
| } |
| |
| int chcr_uld_rx_handler(void *handle, const __be64 *rsp, |
| const struct pkt_gl *pgl) |
| { |
| struct uld_ctx *u_ctx = (struct uld_ctx *)handle; |
| struct chcr_dev *dev = u_ctx->dev; |
| const struct cpl_fw6_pld *rpl = (struct cpl_fw6_pld *)rsp; |
| |
| if (rpl->opcode != CPL_FW6_PLD) { |
| pr_err("Unsupported opcode\n"); |
| return 0; |
| } |
| |
| if (!pgl) |
| work_handlers[rpl->opcode](dev, (unsigned char *)&rsp[1]); |
| else |
| work_handlers[rpl->opcode](dev, pgl->va); |
| return 0; |
| } |
| |
| #ifdef CONFIG_CHELSIO_IPSEC_INLINE |
| int chcr_uld_tx_handler(struct sk_buff *skb, struct net_device *dev) |
| { |
| return chcr_ipsec_xmit(skb, dev); |
| } |
| #endif /* CONFIG_CHELSIO_IPSEC_INLINE */ |
| |
| static int chcr_uld_state_change(void *handle, enum cxgb4_state state) |
| { |
| struct uld_ctx *u_ctx = handle; |
| int ret = 0; |
| |
| switch (state) { |
| case CXGB4_STATE_UP: |
| if (!u_ctx->dev) { |
| ret = chcr_dev_add(u_ctx); |
| if (ret != 0) |
| return ret; |
| } |
| if (atomic_read(&dev_count) == 1) |
| ret = start_crypto(); |
| break; |
| |
| case CXGB4_STATE_DETACH: |
| if (u_ctx->dev) { |
| mutex_lock(&dev_mutex); |
| chcr_dev_remove(u_ctx); |
| mutex_unlock(&dev_mutex); |
| } |
| if (!atomic_read(&dev_count)) |
| stop_crypto(); |
| break; |
| |
| case CXGB4_STATE_START_RECOVERY: |
| case CXGB4_STATE_DOWN: |
| default: |
| break; |
| } |
| return ret; |
| } |
| |
| static int __init chcr_crypto_init(void) |
| { |
| if (cxgb4_register_uld(CXGB4_ULD_CRYPTO, &chcr_uld_info)) |
| pr_err("ULD register fail: No chcr crypto support in cxgb4\n"); |
| |
| return 0; |
| } |
| |
| static void __exit chcr_crypto_exit(void) |
| { |
| struct uld_ctx *u_ctx, *tmp; |
| |
| if (atomic_read(&dev_count)) |
| stop_crypto(); |
| |
| /* Remove all devices from list */ |
| mutex_lock(&dev_mutex); |
| list_for_each_entry_safe(u_ctx, tmp, &uld_ctx_list, entry) { |
| if (u_ctx->dev) |
| chcr_dev_remove(u_ctx); |
| kfree(u_ctx); |
| } |
| mutex_unlock(&dev_mutex); |
| cxgb4_unregister_uld(CXGB4_ULD_CRYPTO); |
| } |
| |
| module_init(chcr_crypto_init); |
| module_exit(chcr_crypto_exit); |
| |
| MODULE_DESCRIPTION("Crypto Co-processor for Chelsio Terminator cards."); |
| MODULE_LICENSE("GPL"); |
| MODULE_AUTHOR("Chelsio Communications"); |
| MODULE_VERSION(DRV_VERSION); |