qed_nvmetcp_if.h (7511B)
1/* SPDX-License-Identifier: (GPL-2.0-only OR BSD-3-Clause) */ 2/* Copyright 2021 Marvell. All rights reserved. */ 3 4#ifndef _QED_NVMETCP_IF_H 5#define _QED_NVMETCP_IF_H 6#include <linux/types.h> 7#include <linux/qed/qed_if.h> 8#include <linux/qed/storage_common.h> 9#include <linux/qed/nvmetcp_common.h> 10 11#define QED_NVMETCP_MAX_IO_SIZE 0x800000 12#define QED_NVMETCP_CMN_HDR_SIZE (sizeof(struct nvme_tcp_hdr)) 13#define QED_NVMETCP_CMD_HDR_SIZE (sizeof(struct nvme_tcp_cmd_pdu)) 14#define QED_NVMETCP_NON_IO_HDR_SIZE ((QED_NVMETCP_CMN_HDR_SIZE + 16)) 15 16typedef int (*nvmetcp_event_cb_t) (void *context, 17 u8 fw_event_code, void *fw_handle); 18 19struct qed_dev_nvmetcp_info { 20 struct qed_dev_info common; 21 u8 port_id; /* Physical port */ 22 u8 num_cqs; 23}; 24 25#define MAX_TID_BLOCKS_NVMETCP (512) 26struct qed_nvmetcp_tid { 27 u32 size; /* In bytes per task */ 28 u32 num_tids_per_block; 29 u8 *blocks[MAX_TID_BLOCKS_NVMETCP]; 30}; 31 32struct qed_nvmetcp_id_params { 33 u8 mac[ETH_ALEN]; 34 u32 ip[4]; 35 u16 port; 36}; 37 38struct qed_nvmetcp_params_offload { 39 /* FW initializations */ 40 dma_addr_t sq_pbl_addr; 41 dma_addr_t nvmetcp_cccid_itid_table_addr; 42 u16 nvmetcp_cccid_max_range; 43 u8 default_cq; 44 45 /* Networking and TCP stack initializations */ 46 struct qed_nvmetcp_id_params src; 47 struct qed_nvmetcp_id_params dst; 48 u32 ka_timeout; 49 u32 ka_interval; 50 u32 max_rt_time; 51 u32 cwnd; 52 u16 mss; 53 u16 vlan_id; 54 bool timestamp_en; 55 bool delayed_ack_en; 56 bool tcp_keep_alive_en; 57 bool ecn_en; 58 u8 ip_version; 59 u8 ka_max_probe_cnt; 60 u8 ttl; 61 u8 tos_or_tc; 62 u8 rcv_wnd_scale; 63}; 64 65struct qed_nvmetcp_params_update { 66 u32 max_io_size; 67 u32 max_recv_pdu_length; 68 u32 max_send_pdu_length; 69 70 /* Placeholder: pfv, cpda, hpda */ 71 72 bool hdr_digest_en; 73 bool data_digest_en; 74}; 75 76struct qed_nvmetcp_cb_ops { 77 struct qed_common_cb_ops common; 78}; 79 80struct nvmetcp_sge { 81 struct regpair sge_addr; /* SGE address */ 82 __le32 sge_len; /* SGE length */ 83 __le32 reserved; 84}; 85 86/* IO path HSI function SGL params */ 87struct storage_sgl_task_params { 88 struct nvmetcp_sge *sgl; 89 struct regpair sgl_phys_addr; 90 u32 total_buffer_size; 91 u16 num_sges; 92 bool small_mid_sge; 93}; 94 95/* IO path HSI function FW task context params */ 96struct nvmetcp_task_params { 97 void *context; /* Output parameter - set/filled by the HSI function */ 98 struct nvmetcp_wqe *sqe; 99 u32 tx_io_size; /* in bytes (Without DIF, if exists) */ 100 u32 rx_io_size; /* in bytes (Without DIF, if exists) */ 101 u16 conn_icid; 102 u16 itid; 103 struct regpair opq; /* qedn_task_ctx address */ 104 u16 host_cccid; 105 u8 cq_rss_number; 106 bool send_write_incapsule; 107}; 108 109/** 110 * struct qed_nvmetcp_ops - qed NVMeTCP operations. 111 * @common: common operations pointer 112 * @ll2: light L2 operations pointer 113 * @fill_dev_info: fills NVMeTCP specific information 114 * @param cdev 115 * @param info 116 * @return 0 on success, otherwise error value. 117 * @register_ops: register nvmetcp operations 118 * @param cdev 119 * @param ops - specified using qed_nvmetcp_cb_ops 120 * @param cookie - driver private 121 * @start: nvmetcp in FW 122 * @param cdev 123 * @param tasks - qed will fill information about tasks 124 * return 0 on success, otherwise error value. 125 * @stop: nvmetcp in FW 126 * @param cdev 127 * return 0 on success, otherwise error value. 128 * @acquire_conn: acquire a new nvmetcp connection 129 * @param cdev 130 * @param handle - qed will fill handle that should be 131 * used henceforth as identifier of the 132 * connection. 133 * @param p_doorbell - qed will fill the address of the 134 * doorbell. 135 * @return 0 on success, otherwise error value. 136 * @release_conn: release a previously acquired nvmetcp connection 137 * @param cdev 138 * @param handle - the connection handle. 139 * @return 0 on success, otherwise error value. 140 * @offload_conn: configures an offloaded connection 141 * @param cdev 142 * @param handle - the connection handle. 143 * @param conn_info - the configuration to use for the 144 * offload. 145 * @return 0 on success, otherwise error value. 146 * @update_conn: updates an offloaded connection 147 * @param cdev 148 * @param handle - the connection handle. 149 * @param conn_info - the configuration to use for the 150 * offload. 151 * @return 0 on success, otherwise error value. 152 * @destroy_conn: stops an offloaded connection 153 * @param cdev 154 * @param handle - the connection handle. 155 * @return 0 on success, otherwise error value. 156 * @clear_sq: clear all task in sq 157 * @param cdev 158 * @param handle - the connection handle. 159 * @return 0 on success, otherwise error value. 160 * @add_src_tcp_port_filter: Add source tcp port filter 161 * @param cdev 162 * @param src_port 163 * @remove_src_tcp_port_filter: Remove source tcp port filter 164 * @param cdev 165 * @param src_port 166 * @add_dst_tcp_port_filter: Add destination tcp port filter 167 * @param cdev 168 * @param dest_port 169 * @remove_dst_tcp_port_filter: Remove destination tcp port filter 170 * @param cdev 171 * @param dest_port 172 * @clear_all_filters: Clear all filters. 173 * @param cdev 174 * @init_read_io: Init read IO. 175 * @task_params 176 * @cmd_pdu_header 177 * @nvme_cmd 178 * @sgl_task_params 179 * @init_write_io: Init write IO. 180 * @task_params 181 * @cmd_pdu_header 182 * @nvme_cmd 183 * @sgl_task_params 184 * @init_icreq_exchange: Exchange ICReq. 185 * @task_params 186 * @init_conn_req_pdu_hdr 187 * @tx_sgl_task_params 188 * @rx_sgl_task_params 189 * @init_task_cleanup: Init task cleanup. 190 * @task_params 191 */ 192struct qed_nvmetcp_ops { 193 const struct qed_common_ops *common; 194 195 const struct qed_ll2_ops *ll2; 196 197 int (*fill_dev_info)(struct qed_dev *cdev, 198 struct qed_dev_nvmetcp_info *info); 199 200 void (*register_ops)(struct qed_dev *cdev, 201 struct qed_nvmetcp_cb_ops *ops, void *cookie); 202 203 int (*start)(struct qed_dev *cdev, 204 struct qed_nvmetcp_tid *tasks, 205 void *event_context, nvmetcp_event_cb_t async_event_cb); 206 207 int (*stop)(struct qed_dev *cdev); 208 209 int (*acquire_conn)(struct qed_dev *cdev, 210 u32 *handle, 211 u32 *fw_cid, void __iomem **p_doorbell); 212 213 int (*release_conn)(struct qed_dev *cdev, u32 handle); 214 215 int (*offload_conn)(struct qed_dev *cdev, 216 u32 handle, 217 struct qed_nvmetcp_params_offload *conn_info); 218 219 int (*update_conn)(struct qed_dev *cdev, 220 u32 handle, 221 struct qed_nvmetcp_params_update *conn_info); 222 223 int (*destroy_conn)(struct qed_dev *cdev, u32 handle, u8 abrt_conn); 224 225 int (*clear_sq)(struct qed_dev *cdev, u32 handle); 226 227 int (*add_src_tcp_port_filter)(struct qed_dev *cdev, u16 src_port); 228 229 void (*remove_src_tcp_port_filter)(struct qed_dev *cdev, u16 src_port); 230 231 int (*add_dst_tcp_port_filter)(struct qed_dev *cdev, u16 dest_port); 232 233 void (*remove_dst_tcp_port_filter)(struct qed_dev *cdev, u16 dest_port); 234 235 void (*clear_all_filters)(struct qed_dev *cdev); 236 237 void (*init_read_io)(struct nvmetcp_task_params *task_params, 238 struct nvme_tcp_cmd_pdu *cmd_pdu_header, 239 struct nvme_command *nvme_cmd, 240 struct storage_sgl_task_params *sgl_task_params); 241 242 void (*init_write_io)(struct nvmetcp_task_params *task_params, 243 struct nvme_tcp_cmd_pdu *cmd_pdu_header, 244 struct nvme_command *nvme_cmd, 245 struct storage_sgl_task_params *sgl_task_params); 246 247 void (*init_icreq_exchange)(struct nvmetcp_task_params *task_params, 248 struct nvme_tcp_icreq_pdu *init_conn_req_pdu_hdr, 249 struct storage_sgl_task_params *tx_sgl_task_params, 250 struct storage_sgl_task_params *rx_sgl_task_params); 251 252 void (*init_task_cleanup)(struct nvmetcp_task_params *task_params); 253}; 254 255const struct qed_nvmetcp_ops *qed_get_nvmetcp_ops(void); 256void qed_put_nvmetcp_ops(void); 257#endif