cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

qed_nvmetcp.c (24702B)


      1// SPDX-License-Identifier: (GPL-2.0-only OR BSD-3-Clause)
      2/* Copyright 2021 Marvell. All rights reserved. */
      3
      4#include <linux/types.h>
      5#include <asm/byteorder.h>
      6#include <asm/param.h>
      7#include <linux/delay.h>
      8#include <linux/dma-mapping.h>
      9#include <linux/etherdevice.h>
     10#include <linux/kernel.h>
     11#include <linux/log2.h>
     12#include <linux/module.h>
     13#include <linux/pci.h>
     14#include <linux/stddef.h>
     15#include <linux/string.h>
     16#include <linux/errno.h>
     17#include <linux/list.h>
     18#include <linux/qed/qed_nvmetcp_if.h>
     19#include "qed.h"
     20#include "qed_cxt.h"
     21#include "qed_dev_api.h"
     22#include "qed_hsi.h"
     23#include "qed_hw.h"
     24#include "qed_int.h"
     25#include "qed_nvmetcp.h"
     26#include "qed_ll2.h"
     27#include "qed_mcp.h"
     28#include "qed_sp.h"
     29#include "qed_reg_addr.h"
     30#include "qed_nvmetcp_fw_funcs.h"
     31
     32static int qed_nvmetcp_async_event(struct qed_hwfn *p_hwfn, u8 fw_event_code,
     33				   u16 echo, union event_ring_data *data,
     34				   u8 fw_return_code)
     35{
     36	if (p_hwfn->p_nvmetcp_info->event_cb) {
     37		struct qed_nvmetcp_info *p_nvmetcp = p_hwfn->p_nvmetcp_info;
     38
     39		return p_nvmetcp->event_cb(p_nvmetcp->event_context,
     40					 fw_event_code, data);
     41	} else {
     42		DP_NOTICE(p_hwfn, "nvmetcp async completion is not set\n");
     43
     44		return -EINVAL;
     45	}
     46}
     47
     48static int qed_sp_nvmetcp_func_start(struct qed_hwfn *p_hwfn,
     49				     enum spq_mode comp_mode,
     50				     struct qed_spq_comp_cb *p_comp_addr,
     51				     void *event_context,
     52				     nvmetcp_event_cb_t async_event_cb)
     53{
     54	struct nvmetcp_init_ramrod_params *p_ramrod = NULL;
     55	struct qed_nvmetcp_pf_params *p_params = NULL;
     56	struct scsi_init_func_queues *p_queue = NULL;
     57	struct nvmetcp_spe_func_init *p_init = NULL;
     58	struct qed_sp_init_data init_data = {};
     59	struct qed_spq_entry *p_ent = NULL;
     60	int rc = 0;
     61	u16 val;
     62	u8 i;
     63
     64	/* Get SPQ entry */
     65	init_data.cid = qed_spq_get_cid(p_hwfn);
     66	init_data.opaque_fid = p_hwfn->hw_info.opaque_fid;
     67	init_data.comp_mode = comp_mode;
     68	init_data.p_comp_data = p_comp_addr;
     69	rc = qed_sp_init_request(p_hwfn, &p_ent,
     70				 NVMETCP_RAMROD_CMD_ID_INIT_FUNC,
     71				 PROTOCOLID_TCP_ULP, &init_data);
     72	if (rc)
     73		return rc;
     74
     75	p_ramrod = &p_ent->ramrod.nvmetcp_init;
     76	p_init = &p_ramrod->nvmetcp_init_spe;
     77	p_params = &p_hwfn->pf_params.nvmetcp_pf_params;
     78	p_queue = &p_init->q_params;
     79	p_init->num_sq_pages_in_ring = p_params->num_sq_pages_in_ring;
     80	p_init->num_r2tq_pages_in_ring = p_params->num_r2tq_pages_in_ring;
     81	p_init->num_uhq_pages_in_ring = p_params->num_uhq_pages_in_ring;
     82	p_init->ll2_rx_queue_id = RESC_START(p_hwfn, QED_LL2_RAM_QUEUE) +
     83					p_params->ll2_ooo_queue_id;
     84	SET_FIELD(p_init->flags, NVMETCP_SPE_FUNC_INIT_NVMETCP_MODE, 1);
     85	p_init->func_params.log_page_size = ilog2(PAGE_SIZE);
     86	p_init->func_params.num_tasks = cpu_to_le16(p_params->num_tasks);
     87	p_init->debug_flags = p_params->debug_mode;
     88	DMA_REGPAIR_LE(p_queue->glbl_q_params_addr,
     89		       p_params->glbl_q_params_addr);
     90	p_queue->cq_num_entries = cpu_to_le16(QED_NVMETCP_FW_CQ_SIZE);
     91	p_queue->num_queues = p_params->num_queues;
     92	val = RESC_START(p_hwfn, QED_CMDQS_CQS);
     93	p_queue->queue_relative_offset = cpu_to_le16((u16)val);
     94	p_queue->cq_sb_pi = p_params->gl_rq_pi;
     95
     96	for (i = 0; i < p_params->num_queues; i++) {
     97		val = qed_get_igu_sb_id(p_hwfn, i);
     98		p_queue->cq_cmdq_sb_num_arr[i] = cpu_to_le16(val);
     99	}
    100
    101	SET_FIELD(p_queue->q_validity,
    102		  SCSI_INIT_FUNC_QUEUES_CMD_VALID, 0);
    103	p_queue->cmdq_num_entries = 0;
    104	p_queue->bdq_resource_id = (u8)RESC_START(p_hwfn, QED_BDQ);
    105	p_ramrod->tcp_init.two_msl_timer = cpu_to_le32(QED_TCP_TWO_MSL_TIMER);
    106	p_ramrod->tcp_init.tx_sws_timer = cpu_to_le16(QED_TCP_SWS_TIMER);
    107	p_init->half_way_close_timeout = cpu_to_le16(QED_TCP_HALF_WAY_CLOSE_TIMEOUT);
    108	p_ramrod->tcp_init.max_fin_rt = QED_TCP_MAX_FIN_RT;
    109	SET_FIELD(p_ramrod->nvmetcp_init_spe.params,
    110		  NVMETCP_SPE_FUNC_INIT_MAX_SYN_RT, QED_TCP_MAX_FIN_RT);
    111	p_hwfn->p_nvmetcp_info->event_context = event_context;
    112	p_hwfn->p_nvmetcp_info->event_cb = async_event_cb;
    113	qed_spq_register_async_cb(p_hwfn, PROTOCOLID_TCP_ULP,
    114				  qed_nvmetcp_async_event);
    115
    116	return qed_spq_post(p_hwfn, p_ent, NULL);
    117}
    118
    119static int qed_sp_nvmetcp_func_stop(struct qed_hwfn *p_hwfn,
    120				    enum spq_mode comp_mode,
    121				    struct qed_spq_comp_cb *p_comp_addr)
    122{
    123	struct qed_spq_entry *p_ent = NULL;
    124	struct qed_sp_init_data init_data;
    125	int rc;
    126
    127	/* Get SPQ entry */
    128	memset(&init_data, 0, sizeof(init_data));
    129	init_data.cid = qed_spq_get_cid(p_hwfn);
    130	init_data.opaque_fid = p_hwfn->hw_info.opaque_fid;
    131	init_data.comp_mode = comp_mode;
    132	init_data.p_comp_data = p_comp_addr;
    133	rc = qed_sp_init_request(p_hwfn, &p_ent,
    134				 NVMETCP_RAMROD_CMD_ID_DESTROY_FUNC,
    135				 PROTOCOLID_TCP_ULP, &init_data);
    136	if (rc)
    137		return rc;
    138
    139	rc = qed_spq_post(p_hwfn, p_ent, NULL);
    140	qed_spq_unregister_async_cb(p_hwfn, PROTOCOLID_TCP_ULP);
    141
    142	return rc;
    143}
    144
    145static int qed_fill_nvmetcp_dev_info(struct qed_dev *cdev,
    146				     struct qed_dev_nvmetcp_info *info)
    147{
    148	struct qed_hwfn *hwfn = QED_AFFIN_HWFN(cdev);
    149	int rc;
    150
    151	memset(info, 0, sizeof(*info));
    152	rc = qed_fill_dev_info(cdev, &info->common);
    153	info->port_id = MFW_PORT(hwfn);
    154	info->num_cqs = FEAT_NUM(hwfn, QED_NVMETCP_CQ);
    155
    156	return rc;
    157}
    158
    159static void qed_register_nvmetcp_ops(struct qed_dev *cdev,
    160				     struct qed_nvmetcp_cb_ops *ops,
    161				     void *cookie)
    162{
    163	cdev->protocol_ops.nvmetcp = ops;
    164	cdev->ops_cookie = cookie;
    165}
    166
    167static int qed_nvmetcp_stop(struct qed_dev *cdev)
    168{
    169	int rc;
    170
    171	if (!(cdev->flags & QED_FLAG_STORAGE_STARTED)) {
    172		DP_NOTICE(cdev, "nvmetcp already stopped\n");
    173
    174		return 0;
    175	}
    176
    177	if (!hash_empty(cdev->connections)) {
    178		DP_NOTICE(cdev,
    179			  "Can't stop nvmetcp - not all connections were returned\n");
    180
    181		return -EINVAL;
    182	}
    183
    184	/* Stop the nvmetcp */
    185	rc = qed_sp_nvmetcp_func_stop(QED_AFFIN_HWFN(cdev), QED_SPQ_MODE_EBLOCK,
    186				      NULL);
    187	cdev->flags &= ~QED_FLAG_STORAGE_STARTED;
    188
    189	return rc;
    190}
    191
    192static int qed_nvmetcp_start(struct qed_dev *cdev,
    193			     struct qed_nvmetcp_tid *tasks,
    194			     void *event_context,
    195			     nvmetcp_event_cb_t async_event_cb)
    196{
    197	struct qed_tid_mem *tid_info;
    198	int rc;
    199
    200	if (cdev->flags & QED_FLAG_STORAGE_STARTED) {
    201		DP_NOTICE(cdev, "nvmetcp already started;\n");
    202
    203		return 0;
    204	}
    205
    206	rc = qed_sp_nvmetcp_func_start(QED_AFFIN_HWFN(cdev),
    207				       QED_SPQ_MODE_EBLOCK, NULL,
    208				       event_context, async_event_cb);
    209	if (rc) {
    210		DP_NOTICE(cdev, "Failed to start nvmetcp\n");
    211
    212		return rc;
    213	}
    214
    215	cdev->flags |= QED_FLAG_STORAGE_STARTED;
    216	hash_init(cdev->connections);
    217
    218	if (!tasks)
    219		return 0;
    220
    221	tid_info = kzalloc(sizeof(*tid_info), GFP_KERNEL);
    222	if (!tid_info) {
    223		qed_nvmetcp_stop(cdev);
    224
    225		return -ENOMEM;
    226	}
    227
    228	rc = qed_cxt_get_tid_mem_info(QED_AFFIN_HWFN(cdev), tid_info);
    229	if (rc) {
    230		DP_NOTICE(cdev, "Failed to gather task information\n");
    231		qed_nvmetcp_stop(cdev);
    232		kfree(tid_info);
    233
    234		return rc;
    235	}
    236
    237	/* Fill task information */
    238	tasks->size = tid_info->tid_size;
    239	tasks->num_tids_per_block = tid_info->num_tids_per_block;
    240	memcpy(tasks->blocks, tid_info->blocks,
    241	       MAX_TID_BLOCKS_NVMETCP * sizeof(u8 *));
    242	kfree(tid_info);
    243
    244	return 0;
    245}
    246
    247static struct qed_hash_nvmetcp_con *qed_nvmetcp_get_hash(struct qed_dev *cdev,
    248							 u32 handle)
    249{
    250	struct qed_hash_nvmetcp_con *hash_con = NULL;
    251
    252	if (!(cdev->flags & QED_FLAG_STORAGE_STARTED))
    253		return NULL;
    254
    255	hash_for_each_possible(cdev->connections, hash_con, node, handle) {
    256		if (hash_con->con->icid == handle)
    257			break;
    258	}
    259
    260	if (!hash_con || hash_con->con->icid != handle)
    261		return NULL;
    262
    263	return hash_con;
    264}
    265
    266static int qed_sp_nvmetcp_conn_offload(struct qed_hwfn *p_hwfn,
    267				       struct qed_nvmetcp_conn *p_conn,
    268				       enum spq_mode comp_mode,
    269				       struct qed_spq_comp_cb *p_comp_addr)
    270{
    271	struct nvmetcp_spe_conn_offload *p_ramrod = NULL;
    272	struct tcp_offload_params_opt2 *p_tcp = NULL;
    273	struct qed_sp_init_data init_data = { 0 };
    274	struct qed_spq_entry *p_ent = NULL;
    275	dma_addr_t r2tq_pbl_addr;
    276	dma_addr_t xhq_pbl_addr;
    277	dma_addr_t uhq_pbl_addr;
    278	u16 physical_q;
    279	int rc = 0;
    280	u8 i;
    281
    282	/* Get SPQ entry */
    283	init_data.cid = p_conn->icid;
    284	init_data.opaque_fid = p_hwfn->hw_info.opaque_fid;
    285	init_data.comp_mode = comp_mode;
    286	init_data.p_comp_data = p_comp_addr;
    287	rc = qed_sp_init_request(p_hwfn, &p_ent,
    288				 NVMETCP_RAMROD_CMD_ID_OFFLOAD_CONN,
    289				 PROTOCOLID_TCP_ULP, &init_data);
    290	if (rc)
    291		return rc;
    292
    293	p_ramrod = &p_ent->ramrod.nvmetcp_conn_offload;
    294
    295	/* Transmission PQ is the first of the PF */
    296	physical_q = qed_get_cm_pq_idx(p_hwfn, PQ_FLAGS_OFLD);
    297	p_conn->physical_q0 = cpu_to_le16(physical_q);
    298	p_ramrod->nvmetcp.physical_q0 = cpu_to_le16(physical_q);
    299
    300	/* nvmetcp Pure-ACK PQ */
    301	physical_q = qed_get_cm_pq_idx(p_hwfn, PQ_FLAGS_ACK);
    302	p_conn->physical_q1 = cpu_to_le16(physical_q);
    303	p_ramrod->nvmetcp.physical_q1 = cpu_to_le16(physical_q);
    304	p_ramrod->conn_id = cpu_to_le16(p_conn->conn_id);
    305	DMA_REGPAIR_LE(p_ramrod->nvmetcp.sq_pbl_addr, p_conn->sq_pbl_addr);
    306	r2tq_pbl_addr = qed_chain_get_pbl_phys(&p_conn->r2tq);
    307	DMA_REGPAIR_LE(p_ramrod->nvmetcp.r2tq_pbl_addr, r2tq_pbl_addr);
    308	xhq_pbl_addr = qed_chain_get_pbl_phys(&p_conn->xhq);
    309	DMA_REGPAIR_LE(p_ramrod->nvmetcp.xhq_pbl_addr, xhq_pbl_addr);
    310	uhq_pbl_addr = qed_chain_get_pbl_phys(&p_conn->uhq);
    311	DMA_REGPAIR_LE(p_ramrod->nvmetcp.uhq_pbl_addr, uhq_pbl_addr);
    312	p_ramrod->nvmetcp.flags = p_conn->offl_flags;
    313	p_ramrod->nvmetcp.default_cq = p_conn->default_cq;
    314	p_ramrod->nvmetcp.initial_ack = 0;
    315	DMA_REGPAIR_LE(p_ramrod->nvmetcp.nvmetcp.cccid_itid_table_addr,
    316		       p_conn->nvmetcp_cccid_itid_table_addr);
    317	p_ramrod->nvmetcp.nvmetcp.cccid_max_range =
    318		 cpu_to_le16(p_conn->nvmetcp_cccid_max_range);
    319	p_tcp = &p_ramrod->tcp;
    320	qed_set_fw_mac_addr(&p_tcp->remote_mac_addr_hi,
    321			    &p_tcp->remote_mac_addr_mid,
    322			    &p_tcp->remote_mac_addr_lo, p_conn->remote_mac);
    323	qed_set_fw_mac_addr(&p_tcp->local_mac_addr_hi,
    324			    &p_tcp->local_mac_addr_mid,
    325			    &p_tcp->local_mac_addr_lo, p_conn->local_mac);
    326	p_tcp->vlan_id = cpu_to_le16(p_conn->vlan_id);
    327	p_tcp->flags = cpu_to_le16(p_conn->tcp_flags);
    328	p_tcp->ip_version = p_conn->ip_version;
    329	if (p_tcp->ip_version == TCP_IPV6) {
    330		for (i = 0; i < 4; i++) {
    331			p_tcp->remote_ip[i] = cpu_to_le32(p_conn->remote_ip[i]);
    332			p_tcp->local_ip[i] = cpu_to_le32(p_conn->local_ip[i]);
    333		}
    334	} else {
    335		p_tcp->remote_ip[0] = cpu_to_le32(p_conn->remote_ip[0]);
    336		p_tcp->local_ip[0] = cpu_to_le32(p_conn->local_ip[0]);
    337	}
    338
    339	p_tcp->flow_label = cpu_to_le32(p_conn->flow_label);
    340	p_tcp->ttl = p_conn->ttl;
    341	p_tcp->tos_or_tc = p_conn->tos_or_tc;
    342	p_tcp->remote_port = cpu_to_le16(p_conn->remote_port);
    343	p_tcp->local_port = cpu_to_le16(p_conn->local_port);
    344	p_tcp->mss = cpu_to_le16(p_conn->mss);
    345	p_tcp->rcv_wnd_scale = p_conn->rcv_wnd_scale;
    346	p_tcp->connect_mode = p_conn->connect_mode;
    347	p_tcp->cwnd = cpu_to_le32(p_conn->cwnd);
    348	p_tcp->ka_max_probe_cnt = p_conn->ka_max_probe_cnt;
    349	p_tcp->ka_timeout = cpu_to_le32(p_conn->ka_timeout);
    350	p_tcp->max_rt_time = cpu_to_le32(p_conn->max_rt_time);
    351	p_tcp->ka_interval = cpu_to_le32(p_conn->ka_interval);
    352
    353	return qed_spq_post(p_hwfn, p_ent, NULL);
    354}
    355
    356static int qed_sp_nvmetcp_conn_update(struct qed_hwfn *p_hwfn,
    357				      struct qed_nvmetcp_conn *p_conn,
    358				      enum spq_mode comp_mode,
    359				      struct qed_spq_comp_cb *p_comp_addr)
    360{
    361	struct nvmetcp_conn_update_ramrod_params *p_ramrod = NULL;
    362	struct qed_spq_entry *p_ent = NULL;
    363	struct qed_sp_init_data init_data;
    364	int rc = -EINVAL;
    365	u32 dval;
    366
    367	/* Get SPQ entry */
    368	memset(&init_data, 0, sizeof(init_data));
    369	init_data.cid = p_conn->icid;
    370	init_data.opaque_fid = p_hwfn->hw_info.opaque_fid;
    371	init_data.comp_mode = comp_mode;
    372	init_data.p_comp_data = p_comp_addr;
    373
    374	rc = qed_sp_init_request(p_hwfn, &p_ent,
    375				 NVMETCP_RAMROD_CMD_ID_UPDATE_CONN,
    376				 PROTOCOLID_TCP_ULP, &init_data);
    377	if (rc)
    378		return rc;
    379
    380	p_ramrod = &p_ent->ramrod.nvmetcp_conn_update;
    381	p_ramrod->conn_id = cpu_to_le16(p_conn->conn_id);
    382	p_ramrod->flags = p_conn->update_flag;
    383	p_ramrod->max_seq_size = cpu_to_le32(p_conn->max_seq_size);
    384	dval = p_conn->max_recv_pdu_length;
    385	p_ramrod->max_recv_pdu_length = cpu_to_le32(dval);
    386	dval = p_conn->max_send_pdu_length;
    387	p_ramrod->max_send_pdu_length = cpu_to_le32(dval);
    388	p_ramrod->first_seq_length = cpu_to_le32(p_conn->first_seq_length);
    389
    390	return qed_spq_post(p_hwfn, p_ent, NULL);
    391}
    392
    393static int qed_sp_nvmetcp_conn_terminate(struct qed_hwfn *p_hwfn,
    394					 struct qed_nvmetcp_conn *p_conn,
    395					 enum spq_mode comp_mode,
    396					 struct qed_spq_comp_cb *p_comp_addr)
    397{
    398	struct nvmetcp_spe_conn_termination *p_ramrod = NULL;
    399	struct qed_spq_entry *p_ent = NULL;
    400	struct qed_sp_init_data init_data;
    401	int rc = -EINVAL;
    402
    403	/* Get SPQ entry */
    404	memset(&init_data, 0, sizeof(init_data));
    405	init_data.cid = p_conn->icid;
    406	init_data.opaque_fid = p_hwfn->hw_info.opaque_fid;
    407	init_data.comp_mode = comp_mode;
    408	init_data.p_comp_data = p_comp_addr;
    409	rc = qed_sp_init_request(p_hwfn, &p_ent,
    410				 NVMETCP_RAMROD_CMD_ID_TERMINATION_CONN,
    411				 PROTOCOLID_TCP_ULP, &init_data);
    412	if (rc)
    413		return rc;
    414
    415	p_ramrod = &p_ent->ramrod.nvmetcp_conn_terminate;
    416	p_ramrod->conn_id = cpu_to_le16(p_conn->conn_id);
    417	p_ramrod->abortive = p_conn->abortive_dsconnect;
    418
    419	return qed_spq_post(p_hwfn, p_ent, NULL);
    420}
    421
    422static int qed_sp_nvmetcp_conn_clear_sq(struct qed_hwfn *p_hwfn,
    423					struct qed_nvmetcp_conn *p_conn,
    424					enum spq_mode comp_mode,
    425					struct qed_spq_comp_cb *p_comp_addr)
    426{
    427	struct qed_spq_entry *p_ent = NULL;
    428	struct qed_sp_init_data init_data;
    429	int rc = -EINVAL;
    430
    431	/* Get SPQ entry */
    432	memset(&init_data, 0, sizeof(init_data));
    433	init_data.cid = p_conn->icid;
    434	init_data.opaque_fid = p_hwfn->hw_info.opaque_fid;
    435	init_data.comp_mode = comp_mode;
    436	init_data.p_comp_data = p_comp_addr;
    437	rc = qed_sp_init_request(p_hwfn, &p_ent,
    438				 NVMETCP_RAMROD_CMD_ID_CLEAR_SQ,
    439				 PROTOCOLID_TCP_ULP, &init_data);
    440	if (rc)
    441		return rc;
    442
    443	return qed_spq_post(p_hwfn, p_ent, NULL);
    444}
    445
    446static void __iomem *qed_nvmetcp_get_db_addr(struct qed_hwfn *p_hwfn, u32 cid)
    447{
    448	return (u8 __iomem *)p_hwfn->doorbells +
    449			     qed_db_addr(cid, DQ_DEMS_LEGACY);
    450}
    451
    452static int qed_nvmetcp_allocate_connection(struct qed_hwfn *p_hwfn,
    453					   struct qed_nvmetcp_conn **p_out_conn)
    454{
    455	struct qed_chain_init_params params = {
    456		.mode		= QED_CHAIN_MODE_PBL,
    457		.intended_use	= QED_CHAIN_USE_TO_CONSUME_PRODUCE,
    458		.cnt_type	= QED_CHAIN_CNT_TYPE_U16,
    459	};
    460	struct qed_nvmetcp_pf_params *p_params = NULL;
    461	struct qed_nvmetcp_conn *p_conn = NULL;
    462	int rc = 0;
    463
    464	/* Try finding a free connection that can be used */
    465	spin_lock_bh(&p_hwfn->p_nvmetcp_info->lock);
    466	if (!list_empty(&p_hwfn->p_nvmetcp_info->free_list))
    467		p_conn = list_first_entry(&p_hwfn->p_nvmetcp_info->free_list,
    468					  struct qed_nvmetcp_conn, list_entry);
    469	if (p_conn) {
    470		list_del(&p_conn->list_entry);
    471		spin_unlock_bh(&p_hwfn->p_nvmetcp_info->lock);
    472		*p_out_conn = p_conn;
    473
    474		return 0;
    475	}
    476	spin_unlock_bh(&p_hwfn->p_nvmetcp_info->lock);
    477
    478	/* Need to allocate a new connection */
    479	p_params = &p_hwfn->pf_params.nvmetcp_pf_params;
    480	p_conn = kzalloc(sizeof(*p_conn), GFP_KERNEL);
    481	if (!p_conn)
    482		return -ENOMEM;
    483
    484	params.num_elems = p_params->num_r2tq_pages_in_ring *
    485			   QED_CHAIN_PAGE_SIZE / sizeof(struct nvmetcp_wqe);
    486	params.elem_size = sizeof(struct nvmetcp_wqe);
    487	rc = qed_chain_alloc(p_hwfn->cdev, &p_conn->r2tq, &params);
    488	if (rc)
    489		goto nomem_r2tq;
    490
    491	params.num_elems = p_params->num_uhq_pages_in_ring *
    492			   QED_CHAIN_PAGE_SIZE / sizeof(struct iscsi_uhqe);
    493	params.elem_size = sizeof(struct iscsi_uhqe);
    494	rc = qed_chain_alloc(p_hwfn->cdev, &p_conn->uhq, &params);
    495	if (rc)
    496		goto nomem_uhq;
    497
    498	params.elem_size = sizeof(struct iscsi_xhqe);
    499	rc = qed_chain_alloc(p_hwfn->cdev, &p_conn->xhq, &params);
    500	if (rc)
    501		goto nomem;
    502
    503	p_conn->free_on_delete = true;
    504	*p_out_conn = p_conn;
    505
    506	return 0;
    507
    508nomem:
    509	qed_chain_free(p_hwfn->cdev, &p_conn->uhq);
    510nomem_uhq:
    511	qed_chain_free(p_hwfn->cdev, &p_conn->r2tq);
    512nomem_r2tq:
    513	kfree(p_conn);
    514
    515	return -ENOMEM;
    516}
    517
    518static int qed_nvmetcp_acquire_connection(struct qed_hwfn *p_hwfn,
    519					  struct qed_nvmetcp_conn **p_out_conn)
    520{
    521	struct qed_nvmetcp_conn *p_conn = NULL;
    522	int rc = 0;
    523	u32 icid;
    524
    525	spin_lock_bh(&p_hwfn->p_nvmetcp_info->lock);
    526	rc = qed_cxt_acquire_cid(p_hwfn, PROTOCOLID_TCP_ULP, &icid);
    527	spin_unlock_bh(&p_hwfn->p_nvmetcp_info->lock);
    528
    529	if (rc)
    530		return rc;
    531
    532	rc = qed_nvmetcp_allocate_connection(p_hwfn, &p_conn);
    533	if (rc) {
    534		spin_lock_bh(&p_hwfn->p_nvmetcp_info->lock);
    535		qed_cxt_release_cid(p_hwfn, icid);
    536		spin_unlock_bh(&p_hwfn->p_nvmetcp_info->lock);
    537
    538		return rc;
    539	}
    540
    541	p_conn->icid = icid;
    542	p_conn->conn_id = (u16)icid;
    543	p_conn->fw_cid = (p_hwfn->hw_info.opaque_fid << 16) | icid;
    544	*p_out_conn = p_conn;
    545
    546	return rc;
    547}
    548
    549static void qed_nvmetcp_release_connection(struct qed_hwfn *p_hwfn,
    550					   struct qed_nvmetcp_conn *p_conn)
    551{
    552	spin_lock_bh(&p_hwfn->p_nvmetcp_info->lock);
    553	list_add_tail(&p_conn->list_entry, &p_hwfn->p_nvmetcp_info->free_list);
    554	qed_cxt_release_cid(p_hwfn, p_conn->icid);
    555	spin_unlock_bh(&p_hwfn->p_nvmetcp_info->lock);
    556}
    557
    558static void qed_nvmetcp_free_connection(struct qed_hwfn *p_hwfn,
    559					struct qed_nvmetcp_conn *p_conn)
    560{
    561	qed_chain_free(p_hwfn->cdev, &p_conn->xhq);
    562	qed_chain_free(p_hwfn->cdev, &p_conn->uhq);
    563	qed_chain_free(p_hwfn->cdev, &p_conn->r2tq);
    564	kfree(p_conn);
    565}
    566
    567int qed_nvmetcp_alloc(struct qed_hwfn *p_hwfn)
    568{
    569	struct qed_nvmetcp_info *p_nvmetcp_info;
    570
    571	p_nvmetcp_info = kzalloc(sizeof(*p_nvmetcp_info), GFP_KERNEL);
    572	if (!p_nvmetcp_info)
    573		return -ENOMEM;
    574
    575	INIT_LIST_HEAD(&p_nvmetcp_info->free_list);
    576	p_hwfn->p_nvmetcp_info = p_nvmetcp_info;
    577
    578	return 0;
    579}
    580
    581void qed_nvmetcp_setup(struct qed_hwfn *p_hwfn)
    582{
    583	spin_lock_init(&p_hwfn->p_nvmetcp_info->lock);
    584}
    585
    586void qed_nvmetcp_free(struct qed_hwfn *p_hwfn)
    587{
    588	struct qed_nvmetcp_conn *p_conn = NULL;
    589
    590	if (!p_hwfn->p_nvmetcp_info)
    591		return;
    592
    593	while (!list_empty(&p_hwfn->p_nvmetcp_info->free_list)) {
    594		p_conn = list_first_entry(&p_hwfn->p_nvmetcp_info->free_list,
    595					  struct qed_nvmetcp_conn, list_entry);
    596		if (p_conn) {
    597			list_del(&p_conn->list_entry);
    598			qed_nvmetcp_free_connection(p_hwfn, p_conn);
    599		}
    600	}
    601
    602	kfree(p_hwfn->p_nvmetcp_info);
    603	p_hwfn->p_nvmetcp_info = NULL;
    604}
    605
    606static int qed_nvmetcp_acquire_conn(struct qed_dev *cdev,
    607				    u32 *handle,
    608				    u32 *fw_cid, void __iomem **p_doorbell)
    609{
    610	struct qed_hash_nvmetcp_con *hash_con;
    611	int rc;
    612
    613	/* Allocate a hashed connection */
    614	hash_con = kzalloc(sizeof(*hash_con), GFP_ATOMIC);
    615	if (!hash_con)
    616		return -ENOMEM;
    617
    618	/* Acquire the connection */
    619	rc = qed_nvmetcp_acquire_connection(QED_AFFIN_HWFN(cdev),
    620					    &hash_con->con);
    621	if (rc) {
    622		DP_NOTICE(cdev, "Failed to acquire Connection\n");
    623		kfree(hash_con);
    624
    625		return rc;
    626	}
    627
    628	/* Added the connection to hash table */
    629	*handle = hash_con->con->icid;
    630	*fw_cid = hash_con->con->fw_cid;
    631	hash_add(cdev->connections, &hash_con->node, *handle);
    632	if (p_doorbell)
    633		*p_doorbell = qed_nvmetcp_get_db_addr(QED_AFFIN_HWFN(cdev),
    634						      *handle);
    635
    636	return 0;
    637}
    638
    639static int qed_nvmetcp_release_conn(struct qed_dev *cdev, u32 handle)
    640{
    641	struct qed_hash_nvmetcp_con *hash_con;
    642
    643	hash_con = qed_nvmetcp_get_hash(cdev, handle);
    644	if (!hash_con) {
    645		DP_NOTICE(cdev, "Failed to find connection for handle %d\n",
    646			  handle);
    647
    648		return -EINVAL;
    649	}
    650
    651	hlist_del(&hash_con->node);
    652	qed_nvmetcp_release_connection(QED_AFFIN_HWFN(cdev), hash_con->con);
    653	kfree(hash_con);
    654
    655	return 0;
    656}
    657
    658static int qed_nvmetcp_offload_conn(struct qed_dev *cdev, u32 handle,
    659				    struct qed_nvmetcp_params_offload *conn_info)
    660{
    661	struct qed_hash_nvmetcp_con *hash_con;
    662	struct qed_nvmetcp_conn *con;
    663
    664	hash_con = qed_nvmetcp_get_hash(cdev, handle);
    665	if (!hash_con) {
    666		DP_NOTICE(cdev, "Failed to find connection for handle %d\n",
    667			  handle);
    668
    669		return -EINVAL;
    670	}
    671
    672	/* Update the connection with information from the params */
    673	con = hash_con->con;
    674
    675	/* FW initializations */
    676	con->layer_code = NVMETCP_SLOW_PATH_LAYER_CODE;
    677	con->sq_pbl_addr = conn_info->sq_pbl_addr;
    678	con->nvmetcp_cccid_max_range = conn_info->nvmetcp_cccid_max_range;
    679	con->nvmetcp_cccid_itid_table_addr = conn_info->nvmetcp_cccid_itid_table_addr;
    680	con->default_cq = conn_info->default_cq;
    681	SET_FIELD(con->offl_flags, NVMETCP_CONN_OFFLOAD_PARAMS_TARGET_MODE, 0);
    682	SET_FIELD(con->offl_flags, NVMETCP_CONN_OFFLOAD_PARAMS_NVMETCP_MODE, 1);
    683	SET_FIELD(con->offl_flags, NVMETCP_CONN_OFFLOAD_PARAMS_TCP_ON_CHIP_1B, 1);
    684
    685	/* Networking and TCP stack initializations */
    686	ether_addr_copy(con->local_mac, conn_info->src.mac);
    687	ether_addr_copy(con->remote_mac, conn_info->dst.mac);
    688	memcpy(con->local_ip, conn_info->src.ip, sizeof(con->local_ip));
    689	memcpy(con->remote_ip, conn_info->dst.ip, sizeof(con->remote_ip));
    690	con->local_port = conn_info->src.port;
    691	con->remote_port = conn_info->dst.port;
    692	con->vlan_id = conn_info->vlan_id;
    693
    694	if (conn_info->timestamp_en)
    695		SET_FIELD(con->tcp_flags, TCP_OFFLOAD_PARAMS_OPT2_TS_EN, 1);
    696
    697	if (conn_info->delayed_ack_en)
    698		SET_FIELD(con->tcp_flags, TCP_OFFLOAD_PARAMS_OPT2_DA_EN, 1);
    699
    700	if (conn_info->tcp_keep_alive_en)
    701		SET_FIELD(con->tcp_flags, TCP_OFFLOAD_PARAMS_OPT2_KA_EN, 1);
    702
    703	if (conn_info->ecn_en)
    704		SET_FIELD(con->tcp_flags, TCP_OFFLOAD_PARAMS_OPT2_ECN_EN, 1);
    705
    706	con->ip_version = conn_info->ip_version;
    707	con->flow_label = QED_TCP_FLOW_LABEL;
    708	con->ka_max_probe_cnt = conn_info->ka_max_probe_cnt;
    709	con->ka_timeout = conn_info->ka_timeout;
    710	con->ka_interval = conn_info->ka_interval;
    711	con->max_rt_time = conn_info->max_rt_time;
    712	con->ttl = conn_info->ttl;
    713	con->tos_or_tc = conn_info->tos_or_tc;
    714	con->mss = conn_info->mss;
    715	con->cwnd = conn_info->cwnd;
    716	con->rcv_wnd_scale = conn_info->rcv_wnd_scale;
    717	con->connect_mode = 0;
    718
    719	return qed_sp_nvmetcp_conn_offload(QED_AFFIN_HWFN(cdev), con,
    720					 QED_SPQ_MODE_EBLOCK, NULL);
    721}
    722
    723static int qed_nvmetcp_update_conn(struct qed_dev *cdev,
    724				   u32 handle,
    725				   struct qed_nvmetcp_params_update *conn_info)
    726{
    727	struct qed_hash_nvmetcp_con *hash_con;
    728	struct qed_nvmetcp_conn *con;
    729
    730	hash_con = qed_nvmetcp_get_hash(cdev, handle);
    731	if (!hash_con) {
    732		DP_NOTICE(cdev, "Failed to find connection for handle %d\n",
    733			  handle);
    734
    735		return -EINVAL;
    736	}
    737
    738	/* Update the connection with information from the params */
    739	con = hash_con->con;
    740	SET_FIELD(con->update_flag,
    741		  ISCSI_CONN_UPDATE_RAMROD_PARAMS_INITIAL_R2T, 0);
    742	SET_FIELD(con->update_flag,
    743		  ISCSI_CONN_UPDATE_RAMROD_PARAMS_IMMEDIATE_DATA, 1);
    744	if (conn_info->hdr_digest_en)
    745		SET_FIELD(con->update_flag, ISCSI_CONN_UPDATE_RAMROD_PARAMS_HD_EN, 1);
    746
    747	if (conn_info->data_digest_en)
    748		SET_FIELD(con->update_flag, ISCSI_CONN_UPDATE_RAMROD_PARAMS_DD_EN, 1);
    749
    750	/* Placeholder - initialize pfv, cpda, hpda */
    751
    752	con->max_seq_size = conn_info->max_io_size;
    753	con->max_recv_pdu_length = conn_info->max_recv_pdu_length;
    754	con->max_send_pdu_length = conn_info->max_send_pdu_length;
    755	con->first_seq_length = conn_info->max_io_size;
    756
    757	return qed_sp_nvmetcp_conn_update(QED_AFFIN_HWFN(cdev), con,
    758					QED_SPQ_MODE_EBLOCK, NULL);
    759}
    760
    761static int qed_nvmetcp_clear_conn_sq(struct qed_dev *cdev, u32 handle)
    762{
    763	struct qed_hash_nvmetcp_con *hash_con;
    764
    765	hash_con = qed_nvmetcp_get_hash(cdev, handle);
    766	if (!hash_con) {
    767		DP_NOTICE(cdev, "Failed to find connection for handle %d\n",
    768			  handle);
    769
    770		return -EINVAL;
    771	}
    772
    773	return qed_sp_nvmetcp_conn_clear_sq(QED_AFFIN_HWFN(cdev), hash_con->con,
    774					    QED_SPQ_MODE_EBLOCK, NULL);
    775}
    776
    777static int qed_nvmetcp_destroy_conn(struct qed_dev *cdev,
    778				    u32 handle, u8 abrt_conn)
    779{
    780	struct qed_hash_nvmetcp_con *hash_con;
    781
    782	hash_con = qed_nvmetcp_get_hash(cdev, handle);
    783	if (!hash_con) {
    784		DP_NOTICE(cdev, "Failed to find connection for handle %d\n",
    785			  handle);
    786
    787		return -EINVAL;
    788	}
    789
    790	hash_con->con->abortive_dsconnect = abrt_conn;
    791
    792	return qed_sp_nvmetcp_conn_terminate(QED_AFFIN_HWFN(cdev), hash_con->con,
    793					   QED_SPQ_MODE_EBLOCK, NULL);
    794}
    795
    796static const struct qed_nvmetcp_ops qed_nvmetcp_ops_pass = {
    797	.common = &qed_common_ops_pass,
    798	.ll2 = &qed_ll2_ops_pass,
    799	.fill_dev_info = &qed_fill_nvmetcp_dev_info,
    800	.register_ops = &qed_register_nvmetcp_ops,
    801	.start = &qed_nvmetcp_start,
    802	.stop = &qed_nvmetcp_stop,
    803	.acquire_conn = &qed_nvmetcp_acquire_conn,
    804	.release_conn = &qed_nvmetcp_release_conn,
    805	.offload_conn = &qed_nvmetcp_offload_conn,
    806	.update_conn = &qed_nvmetcp_update_conn,
    807	.destroy_conn = &qed_nvmetcp_destroy_conn,
    808	.clear_sq = &qed_nvmetcp_clear_conn_sq,
    809	.add_src_tcp_port_filter = &qed_llh_add_src_tcp_port_filter,
    810	.remove_src_tcp_port_filter = &qed_llh_remove_src_tcp_port_filter,
    811	.add_dst_tcp_port_filter = &qed_llh_add_dst_tcp_port_filter,
    812	.remove_dst_tcp_port_filter = &qed_llh_remove_dst_tcp_port_filter,
    813	.clear_all_filters = &qed_llh_clear_all_filters,
    814	.init_read_io = &init_nvmetcp_host_read_task,
    815	.init_write_io = &init_nvmetcp_host_write_task,
    816	.init_icreq_exchange = &init_nvmetcp_init_conn_req_task,
    817	.init_task_cleanup = &init_cleanup_task_nvmetcp
    818};
    819
    820const struct qed_nvmetcp_ops *qed_get_nvmetcp_ops(void)
    821{
    822	return &qed_nvmetcp_ops_pass;
    823}
    824EXPORT_SYMBOL(qed_get_nvmetcp_ops);
    825
    826void qed_put_nvmetcp_ops(void)
    827{
    828}
    829EXPORT_SYMBOL(qed_put_nvmetcp_ops);