1 /* SPDX-License-Identifier: GPL-2.0-only */ 2 /* Copyright (C) 2023 Intel Corporation */ 3 4 #ifndef _VIRTCHNL2_H_ 5 #define _VIRTCHNL2_H_ 6 7 #include <linux/if_ether.h> 8 9 /* All opcodes associated with virtchnl2 are prefixed with virtchnl2 or 10 * VIRTCHNL2. Any future opcodes, offloads/capabilities, structures, 11 * and defines must be prefixed with virtchnl2 or VIRTCHNL2 to avoid confusion. 12 * 13 * PF/VF uses the virtchnl2 interface defined in this header file to communicate 14 * with device Control Plane (CP). Driver and the CP may run on different 15 * platforms with different endianness. To avoid byte order discrepancies, 16 * all the structures in this header follow little-endian format. 17 * 18 * This is an interface definition file where existing enums and their values 19 * must remain unchanged over time, so we specify explicit values for all enums. 20 */ 21 22 /* This macro is used to generate compilation errors if a structure 23 * is not exactly the correct length. 24 */ 25 #define VIRTCHNL2_CHECK_STRUCT_LEN(n, X) \ 26 static_assert((n) == sizeof(struct X)) 27 28 /* New major set of opcodes introduced and so leaving room for 29 * old misc opcodes to be added in future. Also these opcodes may only 30 * be used if both the PF and VF have successfully negotiated the 31 * VIRTCHNL version as 2.0 during VIRTCHNL2_OP_VERSION exchange. 32 */ 33 enum virtchnl2_op { 34 VIRTCHNL2_OP_UNKNOWN = 0, 35 VIRTCHNL2_OP_VERSION = 1, 36 VIRTCHNL2_OP_GET_CAPS = 500, 37 VIRTCHNL2_OP_CREATE_VPORT = 501, 38 VIRTCHNL2_OP_DESTROY_VPORT = 502, 39 VIRTCHNL2_OP_ENABLE_VPORT = 503, 40 VIRTCHNL2_OP_DISABLE_VPORT = 504, 41 VIRTCHNL2_OP_CONFIG_TX_QUEUES = 505, 42 VIRTCHNL2_OP_CONFIG_RX_QUEUES = 506, 43 VIRTCHNL2_OP_ENABLE_QUEUES = 507, 44 VIRTCHNL2_OP_DISABLE_QUEUES = 508, 45 VIRTCHNL2_OP_ADD_QUEUES = 509, 46 VIRTCHNL2_OP_DEL_QUEUES = 510, 47 VIRTCHNL2_OP_MAP_QUEUE_VECTOR = 511, 48 VIRTCHNL2_OP_UNMAP_QUEUE_VECTOR = 512, 49 VIRTCHNL2_OP_GET_RSS_KEY = 513, 50 VIRTCHNL2_OP_SET_RSS_KEY = 514, 51 VIRTCHNL2_OP_GET_RSS_LUT = 515, 52 VIRTCHNL2_OP_SET_RSS_LUT = 516, 53 VIRTCHNL2_OP_GET_RSS_HASH = 517, 54 VIRTCHNL2_OP_SET_RSS_HASH = 518, 55 VIRTCHNL2_OP_SET_SRIOV_VFS = 519, 56 VIRTCHNL2_OP_ALLOC_VECTORS = 520, 57 VIRTCHNL2_OP_DEALLOC_VECTORS = 521, 58 VIRTCHNL2_OP_EVENT = 522, 59 VIRTCHNL2_OP_GET_STATS = 523, 60 VIRTCHNL2_OP_RESET_VF = 524, 61 VIRTCHNL2_OP_GET_EDT_CAPS = 525, 62 VIRTCHNL2_OP_GET_PTYPE_INFO = 526, 63 /* Opcode 527 and 528 are reserved for VIRTCHNL2_OP_GET_PTYPE_ID and 64 * VIRTCHNL2_OP_GET_PTYPE_INFO_RAW. 65 * Opcodes 529, 530, 531, 532 and 533 are reserved. 66 */ 67 VIRTCHNL2_OP_LOOPBACK = 534, 68 VIRTCHNL2_OP_ADD_MAC_ADDR = 535, 69 VIRTCHNL2_OP_DEL_MAC_ADDR = 536, 70 VIRTCHNL2_OP_CONFIG_PROMISCUOUS_MODE = 537, 71 }; 72 73 /** 74 * enum virtchnl2_vport_type - Type of virtual port. 75 * @VIRTCHNL2_VPORT_TYPE_DEFAULT: Default virtual port type. 76 */ 77 enum virtchnl2_vport_type { 78 VIRTCHNL2_VPORT_TYPE_DEFAULT = 0, 79 }; 80 81 /** 82 * enum virtchnl2_queue_model - Type of queue model. 83 * @VIRTCHNL2_QUEUE_MODEL_SINGLE: Single queue model. 84 * @VIRTCHNL2_QUEUE_MODEL_SPLIT: Split queue model. 85 * 86 * In the single queue model, the same transmit descriptor queue is used by 87 * software to post descriptors to hardware and by hardware to post completed 88 * descriptors to software. 89 * Likewise, the same receive descriptor queue is used by hardware to post 90 * completions to software and by software to post buffers to hardware. 91 * 92 * In the split queue model, hardware uses transmit completion queues to post 93 * descriptor/buffer completions to software, while software uses transmit 94 * descriptor queues to post descriptors to hardware. 95 * Likewise, hardware posts descriptor completions to the receive descriptor 96 * queue, while software uses receive buffer queues to post buffers to hardware. 97 */ 98 enum virtchnl2_queue_model { 99 VIRTCHNL2_QUEUE_MODEL_SINGLE = 0, 100 VIRTCHNL2_QUEUE_MODEL_SPLIT = 1, 101 }; 102 103 /* Checksum offload capability flags */ 104 enum virtchnl2_cap_txrx_csum { 105 VIRTCHNL2_CAP_TX_CSUM_L3_IPV4 = BIT(0), 106 VIRTCHNL2_CAP_TX_CSUM_L4_IPV4_TCP = BIT(1), 107 VIRTCHNL2_CAP_TX_CSUM_L4_IPV4_UDP = BIT(2), 108 VIRTCHNL2_CAP_TX_CSUM_L4_IPV4_SCTP = BIT(3), 109 VIRTCHNL2_CAP_TX_CSUM_L4_IPV6_TCP = BIT(4), 110 VIRTCHNL2_CAP_TX_CSUM_L4_IPV6_UDP = BIT(5), 111 VIRTCHNL2_CAP_TX_CSUM_L4_IPV6_SCTP = BIT(6), 112 VIRTCHNL2_CAP_TX_CSUM_GENERIC = BIT(7), 113 VIRTCHNL2_CAP_RX_CSUM_L3_IPV4 = BIT(8), 114 VIRTCHNL2_CAP_RX_CSUM_L4_IPV4_TCP = BIT(9), 115 VIRTCHNL2_CAP_RX_CSUM_L4_IPV4_UDP = BIT(10), 116 VIRTCHNL2_CAP_RX_CSUM_L4_IPV4_SCTP = BIT(11), 117 VIRTCHNL2_CAP_RX_CSUM_L4_IPV6_TCP = BIT(12), 118 VIRTCHNL2_CAP_RX_CSUM_L4_IPV6_UDP = BIT(13), 119 VIRTCHNL2_CAP_RX_CSUM_L4_IPV6_SCTP = BIT(14), 120 VIRTCHNL2_CAP_RX_CSUM_GENERIC = BIT(15), 121 VIRTCHNL2_CAP_TX_CSUM_L3_SINGLE_TUNNEL = BIT(16), 122 VIRTCHNL2_CAP_TX_CSUM_L3_DOUBLE_TUNNEL = BIT(17), 123 VIRTCHNL2_CAP_RX_CSUM_L3_SINGLE_TUNNEL = BIT(18), 124 VIRTCHNL2_CAP_RX_CSUM_L3_DOUBLE_TUNNEL = BIT(19), 125 VIRTCHNL2_CAP_TX_CSUM_L4_SINGLE_TUNNEL = BIT(20), 126 VIRTCHNL2_CAP_TX_CSUM_L4_DOUBLE_TUNNEL = BIT(21), 127 VIRTCHNL2_CAP_RX_CSUM_L4_SINGLE_TUNNEL = BIT(22), 128 VIRTCHNL2_CAP_RX_CSUM_L4_DOUBLE_TUNNEL = BIT(23), 129 }; 130 131 /* Segmentation offload capability flags */ 132 enum virtchnl2_cap_seg { 133 VIRTCHNL2_CAP_SEG_IPV4_TCP = BIT(0), 134 VIRTCHNL2_CAP_SEG_IPV4_UDP = BIT(1), 135 VIRTCHNL2_CAP_SEG_IPV4_SCTP = BIT(2), 136 VIRTCHNL2_CAP_SEG_IPV6_TCP = BIT(3), 137 VIRTCHNL2_CAP_SEG_IPV6_UDP = BIT(4), 138 VIRTCHNL2_CAP_SEG_IPV6_SCTP = BIT(5), 139 VIRTCHNL2_CAP_SEG_GENERIC = BIT(6), 140 VIRTCHNL2_CAP_SEG_TX_SINGLE_TUNNEL = BIT(7), 141 VIRTCHNL2_CAP_SEG_TX_DOUBLE_TUNNEL = BIT(8), 142 }; 143 144 /* Receive Side Scaling Flow type capability flags */ 145 enum virtchnl2_cap_rss { 146 VIRTCHNL2_CAP_RSS_IPV4_TCP = BIT(0), 147 VIRTCHNL2_CAP_RSS_IPV4_UDP = BIT(1), 148 VIRTCHNL2_CAP_RSS_IPV4_SCTP = BIT(2), 149 VIRTCHNL2_CAP_RSS_IPV4_OTHER = BIT(3), 150 VIRTCHNL2_CAP_RSS_IPV6_TCP = BIT(4), 151 VIRTCHNL2_CAP_RSS_IPV6_UDP = BIT(5), 152 VIRTCHNL2_CAP_RSS_IPV6_SCTP = BIT(6), 153 VIRTCHNL2_CAP_RSS_IPV6_OTHER = BIT(7), 154 VIRTCHNL2_CAP_RSS_IPV4_AH = BIT(8), 155 VIRTCHNL2_CAP_RSS_IPV4_ESP = BIT(9), 156 VIRTCHNL2_CAP_RSS_IPV4_AH_ESP = BIT(10), 157 VIRTCHNL2_CAP_RSS_IPV6_AH = BIT(11), 158 VIRTCHNL2_CAP_RSS_IPV6_ESP = BIT(12), 159 VIRTCHNL2_CAP_RSS_IPV6_AH_ESP = BIT(13), 160 }; 161 162 /* Header split capability flags */ 163 enum virtchnl2_cap_rx_hsplit_at { 164 /* for prepended metadata */ 165 VIRTCHNL2_CAP_RX_HSPLIT_AT_L2 = BIT(0), 166 /* all VLANs go into header buffer */ 167 VIRTCHNL2_CAP_RX_HSPLIT_AT_L3 = BIT(1), 168 VIRTCHNL2_CAP_RX_HSPLIT_AT_L4V4 = BIT(2), 169 VIRTCHNL2_CAP_RX_HSPLIT_AT_L4V6 = BIT(3), 170 }; 171 172 /* Receive Side Coalescing offload capability flags */ 173 enum virtchnl2_cap_rsc { 174 VIRTCHNL2_CAP_RSC_IPV4_TCP = BIT(0), 175 VIRTCHNL2_CAP_RSC_IPV4_SCTP = BIT(1), 176 VIRTCHNL2_CAP_RSC_IPV6_TCP = BIT(2), 177 VIRTCHNL2_CAP_RSC_IPV6_SCTP = BIT(3), 178 }; 179 180 /* Other capability flags */ 181 enum virtchnl2_cap_other { 182 VIRTCHNL2_CAP_RDMA = BIT_ULL(0), 183 VIRTCHNL2_CAP_SRIOV = BIT_ULL(1), 184 VIRTCHNL2_CAP_MACFILTER = BIT_ULL(2), 185 VIRTCHNL2_CAP_FLOW_DIRECTOR = BIT_ULL(3), 186 /* Queue based scheduling using split queue model */ 187 VIRTCHNL2_CAP_SPLITQ_QSCHED = BIT_ULL(4), 188 VIRTCHNL2_CAP_CRC = BIT_ULL(5), 189 VIRTCHNL2_CAP_ADQ = BIT_ULL(6), 190 VIRTCHNL2_CAP_WB_ON_ITR = BIT_ULL(7), 191 VIRTCHNL2_CAP_PROMISC = BIT_ULL(8), 192 VIRTCHNL2_CAP_LINK_SPEED = BIT_ULL(9), 193 VIRTCHNL2_CAP_INLINE_IPSEC = BIT_ULL(10), 194 VIRTCHNL2_CAP_LARGE_NUM_QUEUES = BIT_ULL(11), 195 VIRTCHNL2_CAP_VLAN = BIT_ULL(12), 196 VIRTCHNL2_CAP_PTP = BIT_ULL(13), 197 /* EDT: Earliest Departure Time capability used for Timing Wheel */ 198 VIRTCHNL2_CAP_EDT = BIT_ULL(14), 199 VIRTCHNL2_CAP_ADV_RSS = BIT_ULL(15), 200 VIRTCHNL2_CAP_FDIR = BIT_ULL(16), 201 VIRTCHNL2_CAP_RX_FLEX_DESC = BIT_ULL(17), 202 VIRTCHNL2_CAP_PTYPE = BIT_ULL(18), 203 VIRTCHNL2_CAP_LOOPBACK = BIT_ULL(19), 204 /* Other capability 20 is reserved */ 205 206 /* this must be the last capability */ 207 VIRTCHNL2_CAP_OEM = BIT_ULL(63), 208 }; 209 210 /* underlying device type */ 211 enum virtchl2_device_type { 212 VIRTCHNL2_MEV_DEVICE = 0, 213 }; 214 215 /** 216 * enum virtchnl2_txq_sched_mode - Transmit Queue Scheduling Modes. 217 * @VIRTCHNL2_TXQ_SCHED_MODE_QUEUE: Queue mode is the legacy mode i.e. inorder 218 * completions where descriptors and buffers 219 * are completed at the same time. 220 * @VIRTCHNL2_TXQ_SCHED_MODE_FLOW: Flow scheduling mode allows for out of order 221 * packet processing where descriptors are 222 * cleaned in order, but buffers can be 223 * completed out of order. 224 */ 225 enum virtchnl2_txq_sched_mode { 226 VIRTCHNL2_TXQ_SCHED_MODE_QUEUE = 0, 227 VIRTCHNL2_TXQ_SCHED_MODE_FLOW = 1, 228 }; 229 230 /** 231 * enum virtchnl2_rxq_flags - Receive Queue Feature flags. 232 * @VIRTCHNL2_RXQ_RSC: Rx queue RSC flag. 233 * @VIRTCHNL2_RXQ_HDR_SPLIT: Rx queue header split flag. 234 * @VIRTCHNL2_RXQ_IMMEDIATE_WRITE_BACK: When set, packet descriptors are flushed 235 * by hardware immediately after processing 236 * each packet. 237 * @VIRTCHNL2_RX_DESC_SIZE_16BYTE: Rx queue 16 byte descriptor size. 238 * @VIRTCHNL2_RX_DESC_SIZE_32BYTE: Rx queue 32 byte descriptor size. 239 */ 240 enum virtchnl2_rxq_flags { 241 VIRTCHNL2_RXQ_RSC = BIT(0), 242 VIRTCHNL2_RXQ_HDR_SPLIT = BIT(1), 243 VIRTCHNL2_RXQ_IMMEDIATE_WRITE_BACK = BIT(2), 244 VIRTCHNL2_RX_DESC_SIZE_16BYTE = BIT(3), 245 VIRTCHNL2_RX_DESC_SIZE_32BYTE = BIT(4), 246 }; 247 248 /* Type of RSS algorithm */ 249 enum virtchnl2_rss_alg { 250 VIRTCHNL2_RSS_ALG_TOEPLITZ_ASYMMETRIC = 0, 251 VIRTCHNL2_RSS_ALG_R_ASYMMETRIC = 1, 252 VIRTCHNL2_RSS_ALG_TOEPLITZ_SYMMETRIC = 2, 253 VIRTCHNL2_RSS_ALG_XOR_SYMMETRIC = 3, 254 }; 255 256 /* Type of event */ 257 enum virtchnl2_event_codes { 258 VIRTCHNL2_EVENT_UNKNOWN = 0, 259 VIRTCHNL2_EVENT_LINK_CHANGE = 1, 260 /* Event type 2, 3 are reserved */ 261 }; 262 263 /* Transmit and Receive queue types are valid in legacy as well as split queue 264 * models. With Split Queue model, 2 additional types are introduced - 265 * TX_COMPLETION and RX_BUFFER. In split queue model, receive corresponds to 266 * the queue where hardware posts completions. 267 */ 268 enum virtchnl2_queue_type { 269 VIRTCHNL2_QUEUE_TYPE_TX = 0, 270 VIRTCHNL2_QUEUE_TYPE_RX = 1, 271 VIRTCHNL2_QUEUE_TYPE_TX_COMPLETION = 2, 272 VIRTCHNL2_QUEUE_TYPE_RX_BUFFER = 3, 273 VIRTCHNL2_QUEUE_TYPE_CONFIG_TX = 4, 274 VIRTCHNL2_QUEUE_TYPE_CONFIG_RX = 5, 275 /* Queue types 6, 7, 8, 9 are reserved */ 276 VIRTCHNL2_QUEUE_TYPE_MBX_TX = 10, 277 VIRTCHNL2_QUEUE_TYPE_MBX_RX = 11, 278 }; 279 280 /* Interrupt throttling rate index */ 281 enum virtchnl2_itr_idx { 282 VIRTCHNL2_ITR_IDX_0 = 0, 283 VIRTCHNL2_ITR_IDX_1 = 1, 284 }; 285 286 /** 287 * enum virtchnl2_mac_addr_type - MAC address types. 288 * @VIRTCHNL2_MAC_ADDR_PRIMARY: PF/VF driver should set this type for the 289 * primary/device unicast MAC address filter for 290 * VIRTCHNL2_OP_ADD_MAC_ADDR and 291 * VIRTCHNL2_OP_DEL_MAC_ADDR. This allows for the 292 * underlying control plane function to accurately 293 * track the MAC address and for VM/function reset. 294 * 295 * @VIRTCHNL2_MAC_ADDR_EXTRA: PF/VF driver should set this type for any extra 296 * unicast and/or multicast filters that are being 297 * added/deleted via VIRTCHNL2_OP_ADD_MAC_ADDR or 298 * VIRTCHNL2_OP_DEL_MAC_ADDR. 299 */ 300 enum virtchnl2_mac_addr_type { 301 VIRTCHNL2_MAC_ADDR_PRIMARY = 1, 302 VIRTCHNL2_MAC_ADDR_EXTRA = 2, 303 }; 304 305 /* Flags used for promiscuous mode */ 306 enum virtchnl2_promisc_flags { 307 VIRTCHNL2_UNICAST_PROMISC = BIT(0), 308 VIRTCHNL2_MULTICAST_PROMISC = BIT(1), 309 }; 310 311 /* Protocol header type within a packet segment. A segment consists of one or 312 * more protocol headers that make up a logical group of protocol headers. Each 313 * logical group of protocol headers encapsulates or is encapsulated using/by 314 * tunneling or encapsulation protocols for network virtualization. 315 */ 316 enum virtchnl2_proto_hdr_type { 317 /* VIRTCHNL2_PROTO_HDR_ANY is a mandatory protocol id */ 318 VIRTCHNL2_PROTO_HDR_ANY = 0, 319 VIRTCHNL2_PROTO_HDR_PRE_MAC = 1, 320 /* VIRTCHNL2_PROTO_HDR_MAC is a mandatory protocol id */ 321 VIRTCHNL2_PROTO_HDR_MAC = 2, 322 VIRTCHNL2_PROTO_HDR_POST_MAC = 3, 323 VIRTCHNL2_PROTO_HDR_ETHERTYPE = 4, 324 VIRTCHNL2_PROTO_HDR_VLAN = 5, 325 VIRTCHNL2_PROTO_HDR_SVLAN = 6, 326 VIRTCHNL2_PROTO_HDR_CVLAN = 7, 327 VIRTCHNL2_PROTO_HDR_MPLS = 8, 328 VIRTCHNL2_PROTO_HDR_UMPLS = 9, 329 VIRTCHNL2_PROTO_HDR_MMPLS = 10, 330 VIRTCHNL2_PROTO_HDR_PTP = 11, 331 VIRTCHNL2_PROTO_HDR_CTRL = 12, 332 VIRTCHNL2_PROTO_HDR_LLDP = 13, 333 VIRTCHNL2_PROTO_HDR_ARP = 14, 334 VIRTCHNL2_PROTO_HDR_ECP = 15, 335 VIRTCHNL2_PROTO_HDR_EAPOL = 16, 336 VIRTCHNL2_PROTO_HDR_PPPOD = 17, 337 VIRTCHNL2_PROTO_HDR_PPPOE = 18, 338 /* VIRTCHNL2_PROTO_HDR_IPV4 is a mandatory protocol id */ 339 VIRTCHNL2_PROTO_HDR_IPV4 = 19, 340 /* IPv4 and IPv6 Fragment header types are only associated to 341 * VIRTCHNL2_PROTO_HDR_IPV4 and VIRTCHNL2_PROTO_HDR_IPV6 respectively, 342 * cannot be used independently. 343 */ 344 /* VIRTCHNL2_PROTO_HDR_IPV4_FRAG is a mandatory protocol id */ 345 VIRTCHNL2_PROTO_HDR_IPV4_FRAG = 20, 346 /* VIRTCHNL2_PROTO_HDR_IPV6 is a mandatory protocol id */ 347 VIRTCHNL2_PROTO_HDR_IPV6 = 21, 348 /* VIRTCHNL2_PROTO_HDR_IPV6_FRAG is a mandatory protocol id */ 349 VIRTCHNL2_PROTO_HDR_IPV6_FRAG = 22, 350 VIRTCHNL2_PROTO_HDR_IPV6_EH = 23, 351 /* VIRTCHNL2_PROTO_HDR_UDP is a mandatory protocol id */ 352 VIRTCHNL2_PROTO_HDR_UDP = 24, 353 /* VIRTCHNL2_PROTO_HDR_TCP is a mandatory protocol id */ 354 VIRTCHNL2_PROTO_HDR_TCP = 25, 355 /* VIRTCHNL2_PROTO_HDR_SCTP is a mandatory protocol id */ 356 VIRTCHNL2_PROTO_HDR_SCTP = 26, 357 /* VIRTCHNL2_PROTO_HDR_ICMP is a mandatory protocol id */ 358 VIRTCHNL2_PROTO_HDR_ICMP = 27, 359 /* VIRTCHNL2_PROTO_HDR_ICMPV6 is a mandatory protocol id */ 360 VIRTCHNL2_PROTO_HDR_ICMPV6 = 28, 361 VIRTCHNL2_PROTO_HDR_IGMP = 29, 362 VIRTCHNL2_PROTO_HDR_AH = 30, 363 VIRTCHNL2_PROTO_HDR_ESP = 31, 364 VIRTCHNL2_PROTO_HDR_IKE = 32, 365 VIRTCHNL2_PROTO_HDR_NATT_KEEP = 33, 366 /* VIRTCHNL2_PROTO_HDR_PAY is a mandatory protocol id */ 367 VIRTCHNL2_PROTO_HDR_PAY = 34, 368 VIRTCHNL2_PROTO_HDR_L2TPV2 = 35, 369 VIRTCHNL2_PROTO_HDR_L2TPV2_CONTROL = 36, 370 VIRTCHNL2_PROTO_HDR_L2TPV3 = 37, 371 VIRTCHNL2_PROTO_HDR_GTP = 38, 372 VIRTCHNL2_PROTO_HDR_GTP_EH = 39, 373 VIRTCHNL2_PROTO_HDR_GTPCV2 = 40, 374 VIRTCHNL2_PROTO_HDR_GTPC_TEID = 41, 375 VIRTCHNL2_PROTO_HDR_GTPU = 42, 376 VIRTCHNL2_PROTO_HDR_GTPU_UL = 43, 377 VIRTCHNL2_PROTO_HDR_GTPU_DL = 44, 378 VIRTCHNL2_PROTO_HDR_ECPRI = 45, 379 VIRTCHNL2_PROTO_HDR_VRRP = 46, 380 VIRTCHNL2_PROTO_HDR_OSPF = 47, 381 /* VIRTCHNL2_PROTO_HDR_TUN is a mandatory protocol id */ 382 VIRTCHNL2_PROTO_HDR_TUN = 48, 383 VIRTCHNL2_PROTO_HDR_GRE = 49, 384 VIRTCHNL2_PROTO_HDR_NVGRE = 50, 385 VIRTCHNL2_PROTO_HDR_VXLAN = 51, 386 VIRTCHNL2_PROTO_HDR_VXLAN_GPE = 52, 387 VIRTCHNL2_PROTO_HDR_GENEVE = 53, 388 VIRTCHNL2_PROTO_HDR_NSH = 54, 389 VIRTCHNL2_PROTO_HDR_QUIC = 55, 390 VIRTCHNL2_PROTO_HDR_PFCP = 56, 391 VIRTCHNL2_PROTO_HDR_PFCP_NODE = 57, 392 VIRTCHNL2_PROTO_HDR_PFCP_SESSION = 58, 393 VIRTCHNL2_PROTO_HDR_RTP = 59, 394 VIRTCHNL2_PROTO_HDR_ROCE = 60, 395 VIRTCHNL2_PROTO_HDR_ROCEV1 = 61, 396 VIRTCHNL2_PROTO_HDR_ROCEV2 = 62, 397 /* Protocol ids up to 32767 are reserved. 398 * 32768 - 65534 are used for user defined protocol ids. 399 * VIRTCHNL2_PROTO_HDR_NO_PROTO is a mandatory protocol id. 400 */ 401 VIRTCHNL2_PROTO_HDR_NO_PROTO = 65535, 402 }; 403 404 enum virtchl2_version { 405 VIRTCHNL2_VERSION_MINOR_0 = 0, 406 VIRTCHNL2_VERSION_MAJOR_2 = 2, 407 }; 408 409 /** 410 * struct virtchnl2_edt_caps - Get EDT granularity and time horizon. 411 * @tstamp_granularity_ns: Timestamp granularity in nanoseconds. 412 * @time_horizon_ns: Total time window in nanoseconds. 413 * 414 * Associated with VIRTCHNL2_OP_GET_EDT_CAPS. 415 */ 416 struct virtchnl2_edt_caps { 417 __le64 tstamp_granularity_ns; 418 __le64 time_horizon_ns; 419 }; 420 VIRTCHNL2_CHECK_STRUCT_LEN(16, virtchnl2_edt_caps); 421 422 /** 423 * struct virtchnl2_version_info - Version information. 424 * @major: Major version. 425 * @minor: Minor version. 426 * 427 * PF/VF posts its version number to the CP. CP responds with its version number 428 * in the same format, along with a return code. 429 * If there is a major version mismatch, then the PF/VF cannot operate. 430 * If there is a minor version mismatch, then the PF/VF can operate but should 431 * add a warning to the system log. 432 * 433 * This version opcode MUST always be specified as == 1, regardless of other 434 * changes in the API. The CP must always respond to this message without 435 * error regardless of version mismatch. 436 * 437 * Associated with VIRTCHNL2_OP_VERSION. 438 */ 439 struct virtchnl2_version_info { 440 __le32 major; 441 __le32 minor; 442 }; 443 VIRTCHNL2_CHECK_STRUCT_LEN(8, virtchnl2_version_info); 444 445 /** 446 * struct virtchnl2_get_capabilities - Capabilities info. 447 * @csum_caps: See enum virtchnl2_cap_txrx_csum. 448 * @seg_caps: See enum virtchnl2_cap_seg. 449 * @hsplit_caps: See enum virtchnl2_cap_rx_hsplit_at. 450 * @rsc_caps: See enum virtchnl2_cap_rsc. 451 * @rss_caps: See enum virtchnl2_cap_rss. 452 * @other_caps: See enum virtchnl2_cap_other. 453 * @mailbox_dyn_ctl: DYN_CTL register offset and vector id for mailbox 454 * provided by CP. 455 * @mailbox_vector_id: Mailbox vector id. 456 * @num_allocated_vectors: Maximum number of allocated vectors for the device. 457 * @max_rx_q: Maximum number of supported Rx queues. 458 * @max_tx_q: Maximum number of supported Tx queues. 459 * @max_rx_bufq: Maximum number of supported buffer queues. 460 * @max_tx_complq: Maximum number of supported completion queues. 461 * @max_sriov_vfs: The PF sends the maximum VFs it is requesting. The CP 462 * responds with the maximum VFs granted. 463 * @max_vports: Maximum number of vports that can be supported. 464 * @default_num_vports: Default number of vports driver should allocate on load. 465 * @max_tx_hdr_size: Max header length hardware can parse/checksum, in bytes. 466 * @max_sg_bufs_per_tx_pkt: Max number of scatter gather buffers that can be 467 * sent per transmit packet without needing to be 468 * linearized. 469 * @pad: Padding. 470 * @reserved: Reserved. 471 * @device_type: See enum virtchl2_device_type. 472 * @min_sso_packet_len: Min packet length supported by device for single 473 * segment offload. 474 * @max_hdr_buf_per_lso: Max number of header buffers that can be used for 475 * an LSO. 476 * @pad1: Padding for future extensions. 477 * 478 * Dataplane driver sends this message to CP to negotiate capabilities and 479 * provides a virtchnl2_get_capabilities structure with its desired 480 * capabilities, max_sriov_vfs and num_allocated_vectors. 481 * CP responds with a virtchnl2_get_capabilities structure updated 482 * with allowed capabilities and the other fields as below. 483 * If PF sets max_sriov_vfs as 0, CP will respond with max number of VFs 484 * that can be created by this PF. For any other value 'n', CP responds 485 * with max_sriov_vfs set to min(n, x) where x is the max number of VFs 486 * allowed by CP's policy. max_sriov_vfs is not applicable for VFs. 487 * If dataplane driver sets num_allocated_vectors as 0, CP will respond with 1 488 * which is default vector associated with the default mailbox. For any other 489 * value 'n', CP responds with a value <= n based on the CP's policy of 490 * max number of vectors for a PF. 491 * CP will respond with the vector ID of mailbox allocated to the PF in 492 * mailbox_vector_id and the number of itr index registers in itr_idx_map. 493 * It also responds with default number of vports that the dataplane driver 494 * should comeup with in default_num_vports and maximum number of vports that 495 * can be supported in max_vports. 496 * 497 * Associated with VIRTCHNL2_OP_GET_CAPS. 498 */ 499 struct virtchnl2_get_capabilities { 500 __le32 csum_caps; 501 __le32 seg_caps; 502 __le32 hsplit_caps; 503 __le32 rsc_caps; 504 __le64 rss_caps; 505 __le64 other_caps; 506 __le32 mailbox_dyn_ctl; 507 __le16 mailbox_vector_id; 508 __le16 num_allocated_vectors; 509 __le16 max_rx_q; 510 __le16 max_tx_q; 511 __le16 max_rx_bufq; 512 __le16 max_tx_complq; 513 __le16 max_sriov_vfs; 514 __le16 max_vports; 515 __le16 default_num_vports; 516 __le16 max_tx_hdr_size; 517 u8 max_sg_bufs_per_tx_pkt; 518 u8 pad[3]; 519 u8 reserved[4]; 520 __le32 device_type; 521 u8 min_sso_packet_len; 522 u8 max_hdr_buf_per_lso; 523 u8 pad1[10]; 524 }; 525 VIRTCHNL2_CHECK_STRUCT_LEN(80, virtchnl2_get_capabilities); 526 527 /** 528 * struct virtchnl2_queue_reg_chunk - Single queue chunk. 529 * @type: See enum virtchnl2_queue_type. 530 * @start_queue_id: Start Queue ID. 531 * @num_queues: Number of queues in the chunk. 532 * @pad: Padding. 533 * @qtail_reg_start: Queue tail register offset. 534 * @qtail_reg_spacing: Queue tail register spacing. 535 * @pad1: Padding for future extensions. 536 */ 537 struct virtchnl2_queue_reg_chunk { 538 __le32 type; 539 __le32 start_queue_id; 540 __le32 num_queues; 541 __le32 pad; 542 __le64 qtail_reg_start; 543 __le32 qtail_reg_spacing; 544 u8 pad1[4]; 545 }; 546 VIRTCHNL2_CHECK_STRUCT_LEN(32, virtchnl2_queue_reg_chunk); 547 548 /** 549 * struct virtchnl2_queue_reg_chunks - Specify several chunks of contiguous 550 * queues. 551 * @num_chunks: Number of chunks. 552 * @pad: Padding. 553 * @chunks: Chunks of queue info. 554 */ 555 struct virtchnl2_queue_reg_chunks { 556 __le16 num_chunks; 557 u8 pad[6]; 558 struct virtchnl2_queue_reg_chunk chunks[] __counted_by_le(num_chunks); 559 }; 560 VIRTCHNL2_CHECK_STRUCT_LEN(8, virtchnl2_queue_reg_chunks); 561 562 /** 563 * struct virtchnl2_create_vport - Create vport config info. 564 * @vport_type: See enum virtchnl2_vport_type. 565 * @txq_model: See virtchnl2_queue_model. 566 * @rxq_model: See virtchnl2_queue_model. 567 * @num_tx_q: Number of Tx queues. 568 * @num_tx_complq: Valid only if txq_model is split queue. 569 * @num_rx_q: Number of Rx queues. 570 * @num_rx_bufq: Valid only if rxq_model is split queue. 571 * @default_rx_q: Relative receive queue index to be used as default. 572 * @vport_index: Used to align PF and CP in case of default multiple vports, 573 * it is filled by the PF and CP returns the same value, to 574 * enable the driver to support multiple asynchronous parallel 575 * CREATE_VPORT requests and associate a response to a specific 576 * request. 577 * @max_mtu: Max MTU. CP populates this field on response. 578 * @vport_id: Vport id. CP populates this field on response. 579 * @default_mac_addr: Default MAC address. 580 * @pad: Padding. 581 * @rx_desc_ids: See VIRTCHNL2_RX_DESC_IDS definitions. 582 * @tx_desc_ids: See VIRTCHNL2_TX_DESC_IDS definitions. 583 * @pad1: Padding. 584 * @rss_algorithm: RSS algorithm. 585 * @rss_key_size: RSS key size. 586 * @rss_lut_size: RSS LUT size. 587 * @rx_split_pos: See enum virtchnl2_cap_rx_hsplit_at. 588 * @pad2: Padding. 589 * @chunks: Chunks of contiguous queues. 590 * 591 * PF sends this message to CP to create a vport by filling in required 592 * fields of virtchnl2_create_vport structure. 593 * CP responds with the updated virtchnl2_create_vport structure containing the 594 * necessary fields followed by chunks which in turn will have an array of 595 * num_chunks entries of virtchnl2_queue_chunk structures. 596 * 597 * Associated with VIRTCHNL2_OP_CREATE_VPORT. 598 */ 599 struct virtchnl2_create_vport { 600 __le16 vport_type; 601 __le16 txq_model; 602 __le16 rxq_model; 603 __le16 num_tx_q; 604 __le16 num_tx_complq; 605 __le16 num_rx_q; 606 __le16 num_rx_bufq; 607 __le16 default_rx_q; 608 __le16 vport_index; 609 /* CP populates the following fields on response */ 610 __le16 max_mtu; 611 __le32 vport_id; 612 u8 default_mac_addr[ETH_ALEN]; 613 __le16 pad; 614 __le64 rx_desc_ids; 615 __le64 tx_desc_ids; 616 u8 pad1[72]; 617 __le32 rss_algorithm; 618 __le16 rss_key_size; 619 __le16 rss_lut_size; 620 __le32 rx_split_pos; 621 u8 pad2[20]; 622 struct virtchnl2_queue_reg_chunks chunks; 623 }; 624 VIRTCHNL2_CHECK_STRUCT_LEN(160, virtchnl2_create_vport); 625 626 /** 627 * struct virtchnl2_vport - Vport ID info. 628 * @vport_id: Vport id. 629 * @pad: Padding for future extensions. 630 * 631 * PF sends this message to CP to destroy, enable or disable a vport by filling 632 * in the vport_id in virtchnl2_vport structure. 633 * CP responds with the status of the requested operation. 634 * 635 * Associated with VIRTCHNL2_OP_DESTROY_VPORT, VIRTCHNL2_OP_ENABLE_VPORT, 636 * VIRTCHNL2_OP_DISABLE_VPORT. 637 */ 638 struct virtchnl2_vport { 639 __le32 vport_id; 640 u8 pad[4]; 641 }; 642 VIRTCHNL2_CHECK_STRUCT_LEN(8, virtchnl2_vport); 643 644 /** 645 * struct virtchnl2_txq_info - Transmit queue config info 646 * @dma_ring_addr: DMA address. 647 * @type: See enum virtchnl2_queue_type. 648 * @queue_id: Queue ID. 649 * @relative_queue_id: Valid only if queue model is split and type is transmit 650 * queue. Used in many to one mapping of transmit queues to 651 * completion queue. 652 * @model: See enum virtchnl2_queue_model. 653 * @sched_mode: See enum virtchnl2_txq_sched_mode. 654 * @qflags: TX queue feature flags. 655 * @ring_len: Ring length. 656 * @tx_compl_queue_id: Valid only if queue model is split and type is transmit 657 * queue. 658 * @peer_type: Valid only if queue type is VIRTCHNL2_QUEUE_TYPE_MAILBOX_TX 659 * @peer_rx_queue_id: Valid only if queue type is CONFIG_TX and used to deliver 660 * messages for the respective CONFIG_TX queue. 661 * @pad: Padding. 662 * @egress_pasid: Egress PASID info. 663 * @egress_hdr_pasid: Egress HDR passid. 664 * @egress_buf_pasid: Egress buf passid. 665 * @pad1: Padding for future extensions. 666 */ 667 struct virtchnl2_txq_info { 668 __le64 dma_ring_addr; 669 __le32 type; 670 __le32 queue_id; 671 __le16 relative_queue_id; 672 __le16 model; 673 __le16 sched_mode; 674 __le16 qflags; 675 __le16 ring_len; 676 __le16 tx_compl_queue_id; 677 __le16 peer_type; 678 __le16 peer_rx_queue_id; 679 u8 pad[4]; 680 __le32 egress_pasid; 681 __le32 egress_hdr_pasid; 682 __le32 egress_buf_pasid; 683 u8 pad1[8]; 684 }; 685 VIRTCHNL2_CHECK_STRUCT_LEN(56, virtchnl2_txq_info); 686 687 /** 688 * struct virtchnl2_config_tx_queues - TX queue config. 689 * @vport_id: Vport id. 690 * @num_qinfo: Number of virtchnl2_txq_info structs. 691 * @pad: Padding. 692 * @qinfo: Tx queues config info. 693 * 694 * PF sends this message to set up parameters for one or more transmit queues. 695 * This message contains an array of num_qinfo instances of virtchnl2_txq_info 696 * structures. CP configures requested queues and returns a status code. If 697 * num_qinfo specified is greater than the number of queues associated with the 698 * vport, an error is returned and no queues are configured. 699 * 700 * Associated with VIRTCHNL2_OP_CONFIG_TX_QUEUES. 701 */ 702 struct virtchnl2_config_tx_queues { 703 __le32 vport_id; 704 __le16 num_qinfo; 705 u8 pad[10]; 706 struct virtchnl2_txq_info qinfo[] __counted_by_le(num_qinfo); 707 }; 708 VIRTCHNL2_CHECK_STRUCT_LEN(16, virtchnl2_config_tx_queues); 709 710 /** 711 * struct virtchnl2_rxq_info - Receive queue config info. 712 * @desc_ids: See VIRTCHNL2_RX_DESC_IDS definitions. 713 * @dma_ring_addr: See VIRTCHNL2_RX_DESC_IDS definitions. 714 * @type: See enum virtchnl2_queue_type. 715 * @queue_id: Queue id. 716 * @model: See enum virtchnl2_queue_model. 717 * @hdr_buffer_size: Header buffer size. 718 * @data_buffer_size: Data buffer size. 719 * @max_pkt_size: Max packet size. 720 * @ring_len: Ring length. 721 * @buffer_notif_stride: Buffer notification stride in units of 32-descriptors. 722 * This field must be a power of 2. 723 * @pad: Padding. 724 * @dma_head_wb_addr: Applicable only for receive buffer queues. 725 * @qflags: Applicable only for receive completion queues. 726 * See enum virtchnl2_rxq_flags. 727 * @rx_buffer_low_watermark: Rx buffer low watermark. 728 * @rx_bufq1_id: Buffer queue index of the first buffer queue associated with 729 * the Rx queue. Valid only in split queue model. 730 * @rx_bufq2_id: Buffer queue index of the second buffer queue associated with 731 * the Rx queue. Valid only in split queue model. 732 * @bufq2_ena: It indicates if there is a second buffer, rx_bufq2_id is valid 733 * only if this field is set. 734 * @pad1: Padding. 735 * @ingress_pasid: Ingress PASID. 736 * @ingress_hdr_pasid: Ingress PASID header. 737 * @ingress_buf_pasid: Ingress PASID buffer. 738 * @pad2: Padding for future extensions. 739 */ 740 struct virtchnl2_rxq_info { 741 __le64 desc_ids; 742 __le64 dma_ring_addr; 743 __le32 type; 744 __le32 queue_id; 745 __le16 model; 746 __le16 hdr_buffer_size; 747 __le32 data_buffer_size; 748 __le32 max_pkt_size; 749 __le16 ring_len; 750 u8 buffer_notif_stride; 751 u8 pad; 752 __le64 dma_head_wb_addr; 753 __le16 qflags; 754 __le16 rx_buffer_low_watermark; 755 __le16 rx_bufq1_id; 756 __le16 rx_bufq2_id; 757 u8 bufq2_ena; 758 u8 pad1[3]; 759 __le32 ingress_pasid; 760 __le32 ingress_hdr_pasid; 761 __le32 ingress_buf_pasid; 762 u8 pad2[16]; 763 }; 764 VIRTCHNL2_CHECK_STRUCT_LEN(88, virtchnl2_rxq_info); 765 766 /** 767 * struct virtchnl2_config_rx_queues - Rx queues config. 768 * @vport_id: Vport id. 769 * @num_qinfo: Number of instances. 770 * @pad: Padding. 771 * @qinfo: Rx queues config info. 772 * 773 * PF sends this message to set up parameters for one or more receive queues. 774 * This message contains an array of num_qinfo instances of virtchnl2_rxq_info 775 * structures. CP configures requested queues and returns a status code. 776 * If the number of queues specified is greater than the number of queues 777 * associated with the vport, an error is returned and no queues are configured. 778 * 779 * Associated with VIRTCHNL2_OP_CONFIG_RX_QUEUES. 780 */ 781 struct virtchnl2_config_rx_queues { 782 __le32 vport_id; 783 __le16 num_qinfo; 784 u8 pad[18]; 785 struct virtchnl2_rxq_info qinfo[] __counted_by_le(num_qinfo); 786 }; 787 VIRTCHNL2_CHECK_STRUCT_LEN(24, virtchnl2_config_rx_queues); 788 789 /** 790 * struct virtchnl2_add_queues - data for VIRTCHNL2_OP_ADD_QUEUES. 791 * @vport_id: Vport id. 792 * @num_tx_q: Number of Tx qieues. 793 * @num_tx_complq: Number of Tx completion queues. 794 * @num_rx_q: Number of Rx queues. 795 * @num_rx_bufq: Number of Rx buffer queues. 796 * @pad: Padding. 797 * @chunks: Chunks of contiguous queues. 798 * 799 * PF sends this message to request additional transmit/receive queues beyond 800 * the ones that were assigned via CREATE_VPORT request. virtchnl2_add_queues 801 * structure is used to specify the number of each type of queues. 802 * CP responds with the same structure with the actual number of queues assigned 803 * followed by num_chunks of virtchnl2_queue_chunk structures. 804 * 805 * Associated with VIRTCHNL2_OP_ADD_QUEUES. 806 */ 807 struct virtchnl2_add_queues { 808 __le32 vport_id; 809 __le16 num_tx_q; 810 __le16 num_tx_complq; 811 __le16 num_rx_q; 812 __le16 num_rx_bufq; 813 u8 pad[4]; 814 struct virtchnl2_queue_reg_chunks chunks; 815 }; 816 VIRTCHNL2_CHECK_STRUCT_LEN(24, virtchnl2_add_queues); 817 818 /** 819 * struct virtchnl2_vector_chunk - Structure to specify a chunk of contiguous 820 * interrupt vectors. 821 * @start_vector_id: Start vector id. 822 * @start_evv_id: Start EVV id. 823 * @num_vectors: Number of vectors. 824 * @pad: Padding. 825 * @dynctl_reg_start: DYN_CTL register offset. 826 * @dynctl_reg_spacing: register spacing between DYN_CTL registers of 2 827 * consecutive vectors. 828 * @itrn_reg_start: ITRN register offset. 829 * @itrn_reg_spacing: Register spacing between dynctl registers of 2 830 * consecutive vectors. 831 * @itrn_index_spacing: Register spacing between itrn registers of the same 832 * vector where n=0..2. 833 * @pad1: Padding for future extensions. 834 * 835 * Register offsets and spacing provided by CP. 836 * Dynamic control registers are used for enabling/disabling/re-enabling 837 * interrupts and updating interrupt rates in the hotpath. Any changes 838 * to interrupt rates in the dynamic control registers will be reflected 839 * in the interrupt throttling rate registers. 840 * itrn registers are used to update interrupt rates for specific 841 * interrupt indices without modifying the state of the interrupt. 842 */ 843 struct virtchnl2_vector_chunk { 844 __le16 start_vector_id; 845 __le16 start_evv_id; 846 __le16 num_vectors; 847 __le16 pad; 848 __le32 dynctl_reg_start; 849 __le32 dynctl_reg_spacing; 850 __le32 itrn_reg_start; 851 __le32 itrn_reg_spacing; 852 __le32 itrn_index_spacing; 853 u8 pad1[4]; 854 }; 855 VIRTCHNL2_CHECK_STRUCT_LEN(32, virtchnl2_vector_chunk); 856 857 /** 858 * struct virtchnl2_vector_chunks - chunks of contiguous interrupt vectors. 859 * @num_vchunks: number of vector chunks. 860 * @pad: Padding. 861 * @vchunks: Chunks of contiguous vector info. 862 * 863 * PF sends virtchnl2_vector_chunks struct to specify the vectors it is giving 864 * away. CP performs requested action and returns status. 865 * 866 * Associated with VIRTCHNL2_OP_DEALLOC_VECTORS. 867 */ 868 struct virtchnl2_vector_chunks { 869 __le16 num_vchunks; 870 u8 pad[14]; 871 struct virtchnl2_vector_chunk vchunks[] __counted_by_le(num_vchunks); 872 }; 873 VIRTCHNL2_CHECK_STRUCT_LEN(16, virtchnl2_vector_chunks); 874 875 /** 876 * struct virtchnl2_alloc_vectors - vector allocation info. 877 * @num_vectors: Number of vectors. 878 * @pad: Padding. 879 * @vchunks: Chunks of contiguous vector info. 880 * 881 * PF sends this message to request additional interrupt vectors beyond the 882 * ones that were assigned via GET_CAPS request. virtchnl2_alloc_vectors 883 * structure is used to specify the number of vectors requested. CP responds 884 * with the same structure with the actual number of vectors assigned followed 885 * by virtchnl2_vector_chunks structure identifying the vector ids. 886 * 887 * Associated with VIRTCHNL2_OP_ALLOC_VECTORS. 888 */ 889 struct virtchnl2_alloc_vectors { 890 __le16 num_vectors; 891 u8 pad[14]; 892 struct virtchnl2_vector_chunks vchunks; 893 }; 894 VIRTCHNL2_CHECK_STRUCT_LEN(32, virtchnl2_alloc_vectors); 895 896 /** 897 * struct virtchnl2_rss_lut - RSS LUT info. 898 * @vport_id: Vport id. 899 * @lut_entries_start: Start of LUT entries. 900 * @lut_entries: Number of LUT entrties. 901 * @pad: Padding. 902 * @lut: RSS lookup table. 903 * 904 * PF sends this message to get or set RSS lookup table. Only supported if 905 * both PF and CP drivers set the VIRTCHNL2_CAP_RSS bit during configuration 906 * negotiation. 907 * 908 * Associated with VIRTCHNL2_OP_GET_RSS_LUT and VIRTCHNL2_OP_SET_RSS_LUT. 909 */ 910 struct virtchnl2_rss_lut { 911 __le32 vport_id; 912 __le16 lut_entries_start; 913 __le16 lut_entries; 914 u8 pad[4]; 915 __le32 lut[] __counted_by_le(lut_entries); 916 }; 917 VIRTCHNL2_CHECK_STRUCT_LEN(12, virtchnl2_rss_lut); 918 919 /** 920 * struct virtchnl2_rss_hash - RSS hash info. 921 * @ptype_groups: Packet type groups bitmap. 922 * @vport_id: Vport id. 923 * @pad: Padding for future extensions. 924 * 925 * PF sends these messages to get and set the hash filter enable bits for RSS. 926 * By default, the CP sets these to all possible traffic types that the 927 * hardware supports. The PF can query this value if it wants to change the 928 * traffic types that are hashed by the hardware. 929 * Only supported if both PF and CP drivers set the VIRTCHNL2_CAP_RSS bit 930 * during configuration negotiation. 931 * 932 * Associated with VIRTCHNL2_OP_GET_RSS_HASH and VIRTCHNL2_OP_SET_RSS_HASH 933 */ 934 struct virtchnl2_rss_hash { 935 __le64 ptype_groups; 936 __le32 vport_id; 937 u8 pad[4]; 938 }; 939 VIRTCHNL2_CHECK_STRUCT_LEN(16, virtchnl2_rss_hash); 940 941 /** 942 * struct virtchnl2_sriov_vfs_info - VFs info. 943 * @num_vfs: Number of VFs. 944 * @pad: Padding for future extensions. 945 * 946 * This message is used to set number of SRIOV VFs to be created. The actual 947 * allocation of resources for the VFs in terms of vport, queues and interrupts 948 * is done by CP. When this call completes, the IDPF driver calls 949 * pci_enable_sriov to let the OS instantiate the SRIOV PCIE devices. 950 * The number of VFs set to 0 will destroy all the VFs of this function. 951 * 952 * Associated with VIRTCHNL2_OP_SET_SRIOV_VFS. 953 */ 954 struct virtchnl2_sriov_vfs_info { 955 __le16 num_vfs; 956 __le16 pad; 957 }; 958 VIRTCHNL2_CHECK_STRUCT_LEN(4, virtchnl2_sriov_vfs_info); 959 960 /** 961 * struct virtchnl2_ptype - Packet type info. 962 * @ptype_id_10: 10-bit packet type. 963 * @ptype_id_8: 8-bit packet type. 964 * @proto_id_count: Number of protocol ids the packet supports, maximum of 32 965 * protocol ids are supported. 966 * @pad: Padding. 967 * @proto_id: proto_id_count decides the allocation of protocol id array. 968 * See enum virtchnl2_proto_hdr_type. 969 * 970 * Based on the descriptor type the PF supports, CP fills ptype_id_10 or 971 * ptype_id_8 for flex and base descriptor respectively. If ptype_id_10 value 972 * is set to 0xFFFF, PF should consider this ptype as dummy one and it is the 973 * last ptype. 974 */ 975 struct virtchnl2_ptype { 976 __le16 ptype_id_10; 977 u8 ptype_id_8; 978 u8 proto_id_count; 979 __le16 pad; 980 __le16 proto_id[] __counted_by(proto_id_count); 981 } __packed __aligned(2); 982 VIRTCHNL2_CHECK_STRUCT_LEN(6, virtchnl2_ptype); 983 984 /** 985 * struct virtchnl2_get_ptype_info - Packet type info. 986 * @start_ptype_id: Starting ptype ID. 987 * @num_ptypes: Number of packet types from start_ptype_id. 988 * @pad: Padding for future extensions. 989 * 990 * The total number of supported packet types is based on the descriptor type. 991 * For the flex descriptor, it is 1024 (10-bit ptype), and for the base 992 * descriptor, it is 256 (8-bit ptype). Send this message to the CP by 993 * populating the 'start_ptype_id' and the 'num_ptypes'. CP responds with the 994 * 'start_ptype_id', 'num_ptypes', and the array of ptype (virtchnl2_ptype) that 995 * are added at the end of the 'virtchnl2_get_ptype_info' message (Note: There 996 * is no specific field for the ptypes but are added at the end of the 997 * ptype info message. PF/VF is expected to extract the ptypes accordingly. 998 * Reason for doing this is because compiler doesn't allow nested flexible 999 * array fields). 1000 * 1001 * If all the ptypes don't fit into one mailbox buffer, CP splits the 1002 * ptype info into multiple messages, where each message will have its own 1003 * 'start_ptype_id', 'num_ptypes', and the ptype array itself. When CP is done 1004 * updating all the ptype information extracted from the package (the number of 1005 * ptypes extracted might be less than what PF/VF expects), it will append a 1006 * dummy ptype (which has 'ptype_id_10' of 'struct virtchnl2_ptype' as 0xFFFF) 1007 * to the ptype array. 1008 * 1009 * PF/VF is expected to receive multiple VIRTCHNL2_OP_GET_PTYPE_INFO messages. 1010 * 1011 * Associated with VIRTCHNL2_OP_GET_PTYPE_INFO. 1012 */ 1013 struct virtchnl2_get_ptype_info { 1014 __le16 start_ptype_id; 1015 __le16 num_ptypes; 1016 __le32 pad; 1017 }; 1018 VIRTCHNL2_CHECK_STRUCT_LEN(8, virtchnl2_get_ptype_info); 1019 1020 /** 1021 * struct virtchnl2_vport_stats - Vport statistics. 1022 * @vport_id: Vport id. 1023 * @pad: Padding. 1024 * @rx_bytes: Received bytes. 1025 * @rx_unicast: Received unicast packets. 1026 * @rx_multicast: Received multicast packets. 1027 * @rx_broadcast: Received broadcast packets. 1028 * @rx_discards: Discarded packets on receive. 1029 * @rx_errors: Receive errors. 1030 * @rx_unknown_protocol: Unlnown protocol. 1031 * @tx_bytes: Transmitted bytes. 1032 * @tx_unicast: Transmitted unicast packets. 1033 * @tx_multicast: Transmitted multicast packets. 1034 * @tx_broadcast: Transmitted broadcast packets. 1035 * @tx_discards: Discarded packets on transmit. 1036 * @tx_errors: Transmit errors. 1037 * @rx_invalid_frame_length: Packets with invalid frame length. 1038 * @rx_overflow_drop: Packets dropped on buffer overflow. 1039 * 1040 * PF/VF sends this message to CP to get the update stats by specifying the 1041 * vport_id. CP responds with stats in struct virtchnl2_vport_stats. 1042 * 1043 * Associated with VIRTCHNL2_OP_GET_STATS. 1044 */ 1045 struct virtchnl2_vport_stats { 1046 __le32 vport_id; 1047 u8 pad[4]; 1048 __le64 rx_bytes; 1049 __le64 rx_unicast; 1050 __le64 rx_multicast; 1051 __le64 rx_broadcast; 1052 __le64 rx_discards; 1053 __le64 rx_errors; 1054 __le64 rx_unknown_protocol; 1055 __le64 tx_bytes; 1056 __le64 tx_unicast; 1057 __le64 tx_multicast; 1058 __le64 tx_broadcast; 1059 __le64 tx_discards; 1060 __le64 tx_errors; 1061 __le64 rx_invalid_frame_length; 1062 __le64 rx_overflow_drop; 1063 }; 1064 VIRTCHNL2_CHECK_STRUCT_LEN(128, virtchnl2_vport_stats); 1065 1066 /** 1067 * struct virtchnl2_event - Event info. 1068 * @event: Event opcode. See enum virtchnl2_event_codes. 1069 * @link_speed: Link_speed provided in Mbps. 1070 * @vport_id: Vport ID. 1071 * @link_status: Link status. 1072 * @pad: Padding. 1073 * @reserved: Reserved. 1074 * 1075 * CP sends this message to inform the PF/VF driver of events that may affect 1076 * it. No direct response is expected from the driver, though it may generate 1077 * other messages in response to this one. 1078 * 1079 * Associated with VIRTCHNL2_OP_EVENT. 1080 */ 1081 struct virtchnl2_event { 1082 __le32 event; 1083 __le32 link_speed; 1084 __le32 vport_id; 1085 u8 link_status; 1086 u8 pad; 1087 __le16 reserved; 1088 }; 1089 VIRTCHNL2_CHECK_STRUCT_LEN(16, virtchnl2_event); 1090 1091 /** 1092 * struct virtchnl2_rss_key - RSS key info. 1093 * @vport_id: Vport id. 1094 * @key_len: Length of RSS key. 1095 * @pad: Padding. 1096 * @key_flex: RSS hash key, packed bytes. 1097 * PF/VF sends this message to get or set RSS key. Only supported if both 1098 * PF/VF and CP drivers set the VIRTCHNL2_CAP_RSS bit during configuration 1099 * negotiation. 1100 * 1101 * Associated with VIRTCHNL2_OP_GET_RSS_KEY and VIRTCHNL2_OP_SET_RSS_KEY. 1102 */ 1103 struct virtchnl2_rss_key { 1104 __le32 vport_id; 1105 __le16 key_len; 1106 u8 pad; 1107 u8 key_flex[] __counted_by_le(key_len); 1108 } __packed; 1109 VIRTCHNL2_CHECK_STRUCT_LEN(7, virtchnl2_rss_key); 1110 1111 /** 1112 * struct virtchnl2_queue_chunk - chunk of contiguous queues 1113 * @type: See enum virtchnl2_queue_type. 1114 * @start_queue_id: Starting queue id. 1115 * @num_queues: Number of queues. 1116 * @pad: Padding for future extensions. 1117 */ 1118 struct virtchnl2_queue_chunk { 1119 __le32 type; 1120 __le32 start_queue_id; 1121 __le32 num_queues; 1122 u8 pad[4]; 1123 }; 1124 VIRTCHNL2_CHECK_STRUCT_LEN(16, virtchnl2_queue_chunk); 1125 1126 /* struct virtchnl2_queue_chunks - chunks of contiguous queues 1127 * @num_chunks: Number of chunks. 1128 * @pad: Padding. 1129 * @chunks: Chunks of contiguous queues info. 1130 */ 1131 struct virtchnl2_queue_chunks { 1132 __le16 num_chunks; 1133 u8 pad[6]; 1134 struct virtchnl2_queue_chunk chunks[] __counted_by_le(num_chunks); 1135 }; 1136 VIRTCHNL2_CHECK_STRUCT_LEN(8, virtchnl2_queue_chunks); 1137 1138 /** 1139 * struct virtchnl2_del_ena_dis_queues - Enable/disable queues info. 1140 * @vport_id: Vport id. 1141 * @pad: Padding. 1142 * @chunks: Chunks of contiguous queues info. 1143 * 1144 * PF sends these messages to enable, disable or delete queues specified in 1145 * chunks. PF sends virtchnl2_del_ena_dis_queues struct to specify the queues 1146 * to be enabled/disabled/deleted. Also applicable to single queue receive or 1147 * transmit. CP performs requested action and returns status. 1148 * 1149 * Associated with VIRTCHNL2_OP_ENABLE_QUEUES, VIRTCHNL2_OP_DISABLE_QUEUES and 1150 * VIRTCHNL2_OP_DISABLE_QUEUES. 1151 */ 1152 struct virtchnl2_del_ena_dis_queues { 1153 __le32 vport_id; 1154 u8 pad[4]; 1155 struct virtchnl2_queue_chunks chunks; 1156 }; 1157 VIRTCHNL2_CHECK_STRUCT_LEN(16, virtchnl2_del_ena_dis_queues); 1158 1159 /** 1160 * struct virtchnl2_queue_vector - Queue to vector mapping. 1161 * @queue_id: Queue id. 1162 * @vector_id: Vector id. 1163 * @pad: Padding. 1164 * @itr_idx: See enum virtchnl2_itr_idx. 1165 * @queue_type: See enum virtchnl2_queue_type. 1166 * @pad1: Padding for future extensions. 1167 */ 1168 struct virtchnl2_queue_vector { 1169 __le32 queue_id; 1170 __le16 vector_id; 1171 u8 pad[2]; 1172 __le32 itr_idx; 1173 __le32 queue_type; 1174 u8 pad1[8]; 1175 }; 1176 VIRTCHNL2_CHECK_STRUCT_LEN(24, virtchnl2_queue_vector); 1177 1178 /** 1179 * struct virtchnl2_queue_vector_maps - Map/unmap queues info. 1180 * @vport_id: Vport id. 1181 * @num_qv_maps: Number of queue vector maps. 1182 * @pad: Padding. 1183 * @qv_maps: Queue to vector maps. 1184 * 1185 * PF sends this message to map or unmap queues to vectors and interrupt 1186 * throttling rate index registers. External data buffer contains 1187 * virtchnl2_queue_vector_maps structure that contains num_qv_maps of 1188 * virtchnl2_queue_vector structures. CP maps the requested queue vector maps 1189 * after validating the queue and vector ids and returns a status code. 1190 * 1191 * Associated with VIRTCHNL2_OP_MAP_QUEUE_VECTOR and 1192 * VIRTCHNL2_OP_UNMAP_QUEUE_VECTOR. 1193 */ 1194 struct virtchnl2_queue_vector_maps { 1195 __le32 vport_id; 1196 __le16 num_qv_maps; 1197 u8 pad[10]; 1198 struct virtchnl2_queue_vector qv_maps[] __counted_by_le(num_qv_maps); 1199 }; 1200 VIRTCHNL2_CHECK_STRUCT_LEN(16, virtchnl2_queue_vector_maps); 1201 1202 /** 1203 * struct virtchnl2_loopback - Loopback info. 1204 * @vport_id: Vport id. 1205 * @enable: Enable/disable. 1206 * @pad: Padding for future extensions. 1207 * 1208 * PF/VF sends this message to transition to/from the loopback state. Setting 1209 * the 'enable' to 1 enables the loopback state and setting 'enable' to 0 1210 * disables it. CP configures the state to loopback and returns status. 1211 * 1212 * Associated with VIRTCHNL2_OP_LOOPBACK. 1213 */ 1214 struct virtchnl2_loopback { 1215 __le32 vport_id; 1216 u8 enable; 1217 u8 pad[3]; 1218 }; 1219 VIRTCHNL2_CHECK_STRUCT_LEN(8, virtchnl2_loopback); 1220 1221 /* struct virtchnl2_mac_addr - MAC address info. 1222 * @addr: MAC address. 1223 * @type: MAC type. See enum virtchnl2_mac_addr_type. 1224 * @pad: Padding for future extensions. 1225 */ 1226 struct virtchnl2_mac_addr { 1227 u8 addr[ETH_ALEN]; 1228 u8 type; 1229 u8 pad; 1230 }; 1231 VIRTCHNL2_CHECK_STRUCT_LEN(8, virtchnl2_mac_addr); 1232 1233 /** 1234 * struct virtchnl2_mac_addr_list - List of MAC addresses. 1235 * @vport_id: Vport id. 1236 * @num_mac_addr: Number of MAC addresses. 1237 * @pad: Padding. 1238 * @mac_addr_list: List with MAC address info. 1239 * 1240 * PF/VF driver uses this structure to send list of MAC addresses to be 1241 * added/deleted to the CP where as CP performs the action and returns the 1242 * status. 1243 * 1244 * Associated with VIRTCHNL2_OP_ADD_MAC_ADDR and VIRTCHNL2_OP_DEL_MAC_ADDR. 1245 */ 1246 struct virtchnl2_mac_addr_list { 1247 __le32 vport_id; 1248 __le16 num_mac_addr; 1249 u8 pad[2]; 1250 struct virtchnl2_mac_addr mac_addr_list[] __counted_by_le(num_mac_addr); 1251 }; 1252 VIRTCHNL2_CHECK_STRUCT_LEN(8, virtchnl2_mac_addr_list); 1253 1254 /** 1255 * struct virtchnl2_promisc_info - Promisc type info. 1256 * @vport_id: Vport id. 1257 * @flags: See enum virtchnl2_promisc_flags. 1258 * @pad: Padding for future extensions. 1259 * 1260 * PF/VF sends vport id and flags to the CP where as CP performs the action 1261 * and returns the status. 1262 * 1263 * Associated with VIRTCHNL2_OP_CONFIG_PROMISCUOUS_MODE. 1264 */ 1265 struct virtchnl2_promisc_info { 1266 __le32 vport_id; 1267 /* See VIRTCHNL2_PROMISC_FLAGS definitions */ 1268 __le16 flags; 1269 u8 pad[2]; 1270 }; 1271 VIRTCHNL2_CHECK_STRUCT_LEN(8, virtchnl2_promisc_info); 1272 1273 #endif /* _VIRTCHNL_2_H_ */ 1274