/* $NetBSD: if_mcx.c,v 1.27 2024/04/11 10:42:42 andvar Exp $ */ /* $OpenBSD: if_mcx.c,v 1.101 2021/06/02 19:16:11 patrick Exp $ */ /* * Copyright (c) 2017 David Gwynne * Copyright (c) 2019 Jonathan Matthew * * Permission to use, copy, modify, and distribute this software for any * purpose with or without fee is hereby granted, provided that the above * copyright notice and this permission notice appear in all copies. * * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. */ #ifdef _KERNEL_OPT #include "opt_net_mpsafe.h" #endif #include __KERNEL_RCSID(0, "$NetBSD: if_mcx.c,v 1.27 2024/04/11 10:42:42 andvar Exp $"); #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include /* TODO: Port kstat key/value stuff to evcnt/sysmon */ #define NKSTAT 0 /* XXX This driver is not yet MP-safe; don't claim to be! */ /* #ifdef NET_MPSAFE */ /* #define MCX_MPSAFE 1 */ /* #define CALLOUT_FLAGS CALLOUT_MPSAFE */ /* #else */ #define CALLOUT_FLAGS 0 /* #endif */ #define MCX_TXQ_NUM 2048 #define BUS_DMASYNC_PRERW (BUS_DMASYNC_PREREAD|BUS_DMASYNC_PREWRITE) #define BUS_DMASYNC_POSTRW (BUS_DMASYNC_POSTREAD|BUS_DMASYNC_POSTWRITE) #define MCX_HCA_BAR PCI_MAPREG_START /* BAR 0 */ #define MCX_FW_VER 0x0000 #define MCX_FW_VER_MAJOR(_v) ((_v) & 0xffff) #define MCX_FW_VER_MINOR(_v) ((_v) >> 16) #define MCX_CMDIF_FW_SUBVER 0x0004 #define MCX_FW_VER_SUBMINOR(_v) ((_v) & 0xffff) #define MCX_CMDIF(_v) ((_v) >> 16) #define MCX_ISSI 1 /* as per the PRM */ #define MCX_CMD_IF_SUPPORTED 5 #define MCX_HARDMTU 9500 #define MCX_PAGE_SHIFT 12 #define MCX_PAGE_SIZE (1 << MCX_PAGE_SHIFT) /* queue sizes */ #define MCX_LOG_EQ_SIZE 7 #define MCX_LOG_CQ_SIZE 12 #define MCX_LOG_RQ_SIZE 10 #define MCX_LOG_SQ_SIZE 11 #define MCX_MAX_QUEUES 16 /* completion event moderation - about 10khz, or 90% of the cq */ #define MCX_CQ_MOD_PERIOD 50 #define MCX_CQ_MOD_COUNTER \ (((1 << (MCX_LOG_CQ_SIZE - 1)) * 9) / 10) #define MCX_LOG_SQ_ENTRY_SIZE 6 #define MCX_SQ_ENTRY_MAX_SLOTS 4 #define MCX_SQ_SEGS_PER_SLOT \ (sizeof(struct mcx_sq_entry) / sizeof(struct mcx_sq_entry_seg)) #define MCX_SQ_MAX_SEGMENTS \ 1 + ((MCX_SQ_ENTRY_MAX_SLOTS-1) * MCX_SQ_SEGS_PER_SLOT) #define MCX_LOG_FLOW_TABLE_SIZE 5 #define MCX_NUM_STATIC_FLOWS 4 /* promisc, allmulti, ucast, bcast */ #define MCX_NUM_MCAST_FLOWS \ ((1 << MCX_LOG_FLOW_TABLE_SIZE) - MCX_NUM_STATIC_FLOWS) #define MCX_SQ_INLINE_SIZE 18 CTASSERT(ETHER_HDR_LEN + ETHER_VLAN_ENCAP_LEN == MCX_SQ_INLINE_SIZE); /* doorbell offsets */ #define MCX_DOORBELL_AREA_SIZE MCX_PAGE_SIZE #define MCX_CQ_DOORBELL_BASE 0 #define MCX_CQ_DOORBELL_STRIDE 64 #define MCX_WQ_DOORBELL_BASE MCX_PAGE_SIZE/2 #define MCX_WQ_DOORBELL_STRIDE 64 /* make sure the doorbells fit */ CTASSERT(MCX_MAX_QUEUES * MCX_CQ_DOORBELL_STRIDE < MCX_WQ_DOORBELL_BASE); CTASSERT(MCX_MAX_QUEUES * MCX_WQ_DOORBELL_STRIDE < MCX_DOORBELL_AREA_SIZE - MCX_WQ_DOORBELL_BASE); #define MCX_WQ_DOORBELL_MASK 0xffff /* uar registers */ #define MCX_UAR_CQ_DOORBELL 0x20 #define MCX_UAR_EQ_DOORBELL_ARM 0x40 #define MCX_UAR_EQ_DOORBELL 0x48 #define MCX_UAR_BF 0x800 #define MCX_CMDQ_ADDR_HI 0x0010 #define MCX_CMDQ_ADDR_LO 0x0014 #define MCX_CMDQ_ADDR_NMASK 0xfff #define MCX_CMDQ_LOG_SIZE(_v) ((_v) >> 4 & 0xf) #define MCX_CMDQ_LOG_STRIDE(_v) ((_v) >> 0 & 0xf) #define MCX_CMDQ_INTERFACE_MASK (0x3 << 8) #define MCX_CMDQ_INTERFACE_FULL_DRIVER (0x0 << 8) #define MCX_CMDQ_INTERFACE_DISABLED (0x1 << 8) #define MCX_CMDQ_DOORBELL 0x0018 #define MCX_STATE 0x01fc #define MCX_STATE_MASK (1U << 31) #define MCX_STATE_INITIALIZING (1 << 31) #define MCX_STATE_READY (0 << 31) #define MCX_STATE_INTERFACE_MASK (0x3 << 24) #define MCX_STATE_INTERFACE_FULL_DRIVER (0x0 << 24) #define MCX_STATE_INTERFACE_DISABLED (0x1 << 24) #define MCX_INTERNAL_TIMER 0x1000 #define MCX_INTERNAL_TIMER_H 0x1000 #define MCX_INTERNAL_TIMER_L 0x1004 #define MCX_CLEAR_INT 0x100c #define MCX_REG_OP_WRITE 0 #define MCX_REG_OP_READ 1 #define MCX_REG_PMLP 0x5002 #define MCX_REG_PMTU 0x5003 #define MCX_REG_PTYS 0x5004 #define MCX_REG_PAOS 0x5006 #define MCX_REG_PFCC 0x5007 #define MCX_REG_PPCNT 0x5008 #define MCX_REG_MTCAP 0x9009 /* mgmt temp capabilities */ #define MCX_REG_MTMP 0x900a /* mgmt temp */ #define MCX_REG_MCIA 0x9014 #define MCX_REG_MCAM 0x907f #define MCX_ETHER_CAP_SGMII 0 #define MCX_ETHER_CAP_1000_KX 1 #define MCX_ETHER_CAP_10G_CX4 2 #define MCX_ETHER_CAP_10G_KX4 3 #define MCX_ETHER_CAP_10G_KR 4 #define MCX_ETHER_CAP_20G_KR2 5 #define MCX_ETHER_CAP_40G_CR4 6 #define MCX_ETHER_CAP_40G_KR4 7 #define MCX_ETHER_CAP_56G_R4 8 #define MCX_ETHER_CAP_10G_CR 12 #define MCX_ETHER_CAP_10G_SR 13 #define MCX_ETHER_CAP_10G_LR 14 #define MCX_ETHER_CAP_40G_SR4 15 #define MCX_ETHER_CAP_40G_LR4 16 #define MCX_ETHER_CAP_50G_SR2 18 #define MCX_ETHER_CAP_100G_CR4 20 #define MCX_ETHER_CAP_100G_SR4 21 #define MCX_ETHER_CAP_100G_KR4 22 #define MCX_ETHER_CAP_100G_LR4 23 #define MCX_ETHER_CAP_100_TX 24 #define MCX_ETHER_CAP_1000_T 25 #define MCX_ETHER_CAP_10G_T 26 #define MCX_ETHER_CAP_25G_CR 27 #define MCX_ETHER_CAP_25G_KR 28 #define MCX_ETHER_CAP_25G_SR 29 #define MCX_ETHER_CAP_50G_CR2 30 #define MCX_ETHER_CAP_50G_KR2 31 #define MCX_MAX_CQE 32 #define MCX_CMD_QUERY_HCA_CAP 0x100 #define MCX_CMD_QUERY_ADAPTER 0x101 #define MCX_CMD_INIT_HCA 0x102 #define MCX_CMD_TEARDOWN_HCA 0x103 #define MCX_CMD_ENABLE_HCA 0x104 #define MCX_CMD_DISABLE_HCA 0x105 #define MCX_CMD_QUERY_PAGES 0x107 #define MCX_CMD_MANAGE_PAGES 0x108 #define MCX_CMD_SET_HCA_CAP 0x109 #define MCX_CMD_QUERY_ISSI 0x10a #define MCX_CMD_SET_ISSI 0x10b #define MCX_CMD_SET_DRIVER_VERSION 0x10d #define MCX_CMD_QUERY_SPECIAL_CONTEXTS 0x203 #define MCX_CMD_CREATE_EQ 0x301 #define MCX_CMD_DESTROY_EQ 0x302 #define MCX_CMD_QUERY_EQ 0x303 #define MCX_CMD_CREATE_CQ 0x400 #define MCX_CMD_DESTROY_CQ 0x401 #define MCX_CMD_QUERY_CQ 0x402 #define MCX_CMD_QUERY_NIC_VPORT_CONTEXT 0x754 #define MCX_CMD_MODIFY_NIC_VPORT_CONTEXT \ 0x755 #define MCX_CMD_QUERY_VPORT_COUNTERS 0x770 #define MCX_CMD_ALLOC_PD 0x800 #define MCX_CMD_ALLOC_UAR 0x802 #define MCX_CMD_ACCESS_REG 0x805 #define MCX_CMD_ALLOC_TRANSPORT_DOMAIN 0x816 #define MCX_CMD_CREATE_TIR 0x900 #define MCX_CMD_DESTROY_TIR 0x902 #define MCX_CMD_CREATE_SQ 0x904 #define MCX_CMD_MODIFY_SQ 0x905 #define MCX_CMD_DESTROY_SQ 0x906 #define MCX_CMD_QUERY_SQ 0x907 #define MCX_CMD_CREATE_RQ 0x908 #define MCX_CMD_MODIFY_RQ 0x909 #define MCX_CMD_DESTROY_RQ 0x90a #define MCX_CMD_QUERY_RQ 0x90b #define MCX_CMD_CREATE_TIS 0x912 #define MCX_CMD_DESTROY_TIS 0x914 #define MCX_CMD_CREATE_RQT 0x916 #define MCX_CMD_DESTROY_RQT 0x918 #define MCX_CMD_SET_FLOW_TABLE_ROOT 0x92f #define MCX_CMD_CREATE_FLOW_TABLE 0x930 #define MCX_CMD_DESTROY_FLOW_TABLE 0x931 #define MCX_CMD_QUERY_FLOW_TABLE 0x932 #define MCX_CMD_CREATE_FLOW_GROUP 0x933 #define MCX_CMD_DESTROY_FLOW_GROUP 0x934 #define MCX_CMD_QUERY_FLOW_GROUP 0x935 #define MCX_CMD_SET_FLOW_TABLE_ENTRY 0x936 #define MCX_CMD_QUERY_FLOW_TABLE_ENTRY 0x937 #define MCX_CMD_DELETE_FLOW_TABLE_ENTRY 0x938 #define MCX_CMD_ALLOC_FLOW_COUNTER 0x939 #define MCX_CMD_QUERY_FLOW_COUNTER 0x93b #define MCX_QUEUE_STATE_RST 0 #define MCX_QUEUE_STATE_RDY 1 #define MCX_QUEUE_STATE_ERR 3 #define MCX_FLOW_TABLE_TYPE_RX 0 #define MCX_FLOW_TABLE_TYPE_TX 1 #define MCX_CMDQ_INLINE_DATASIZE 16 struct mcx_cmdq_entry { uint8_t cq_type; #define MCX_CMDQ_TYPE_PCIE 0x7 uint8_t cq_reserved0[3]; uint32_t cq_input_length; uint64_t cq_input_ptr; uint8_t cq_input_data[MCX_CMDQ_INLINE_DATASIZE]; uint8_t cq_output_data[MCX_CMDQ_INLINE_DATASIZE]; uint64_t cq_output_ptr; uint32_t cq_output_length; uint8_t cq_token; uint8_t cq_signature; uint8_t cq_reserved1[1]; uint8_t cq_status; #define MCX_CQ_STATUS_SHIFT 1 #define MCX_CQ_STATUS_MASK (0x7f << MCX_CQ_STATUS_SHIFT) #define MCX_CQ_STATUS_OK (0x00 << MCX_CQ_STATUS_SHIFT) #define MCX_CQ_STATUS_INT_ERR (0x01 << MCX_CQ_STATUS_SHIFT) #define MCX_CQ_STATUS_BAD_OPCODE (0x02 << MCX_CQ_STATUS_SHIFT) #define MCX_CQ_STATUS_BAD_PARAM (0x03 << MCX_CQ_STATUS_SHIFT) #define MCX_CQ_STATUS_BAD_SYS_STATE (0x04 << MCX_CQ_STATUS_SHIFT) #define MCX_CQ_STATUS_BAD_RESOURCE (0x05 << MCX_CQ_STATUS_SHIFT) #define MCX_CQ_STATUS_RESOURCE_BUSY (0x06 << MCX_CQ_STATUS_SHIFT) #define MCX_CQ_STATUS_EXCEED_LIM (0x08 << MCX_CQ_STATUS_SHIFT) #define MCX_CQ_STATUS_BAD_RES_STATE (0x09 << MCX_CQ_STATUS_SHIFT) #define MCX_CQ_STATUS_BAD_INDEX (0x0a << MCX_CQ_STATUS_SHIFT) #define MCX_CQ_STATUS_NO_RESOURCES (0x0f << MCX_CQ_STATUS_SHIFT) #define MCX_CQ_STATUS_BAD_INPUT_LEN (0x50 << MCX_CQ_STATUS_SHIFT) #define MCX_CQ_STATUS_BAD_OUTPUT_LEN (0x51 << MCX_CQ_STATUS_SHIFT) #define MCX_CQ_STATUS_BAD_RESOURCE_STATE \ (0x10 << MCX_CQ_STATUS_SHIFT) #define MCX_CQ_STATUS_BAD_SIZE (0x40 << MCX_CQ_STATUS_SHIFT) #define MCX_CQ_STATUS_OWN_MASK 0x1 #define MCX_CQ_STATUS_OWN_SW 0x0 #define MCX_CQ_STATUS_OWN_HW 0x1 } __packed __aligned(8); #define MCX_CMDQ_MAILBOX_DATASIZE 512 struct mcx_cmdq_mailbox { uint8_t mb_data[MCX_CMDQ_MAILBOX_DATASIZE]; uint8_t mb_reserved0[48]; uint64_t mb_next_ptr; uint32_t mb_block_number; uint8_t mb_reserved1[1]; uint8_t mb_token; uint8_t mb_ctrl_signature; uint8_t mb_signature; } __packed __aligned(8); #define MCX_CMDQ_MAILBOX_ALIGN (1 << 10) #define MCX_CMDQ_MAILBOX_SIZE roundup(sizeof(struct mcx_cmdq_mailbox), \ MCX_CMDQ_MAILBOX_ALIGN) /* * command mailbox structures */ struct mcx_cmd_enable_hca_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[2]; uint16_t cmd_function_id; uint8_t cmd_reserved2[4]; } __packed __aligned(4); struct mcx_cmd_enable_hca_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[4]; } __packed __aligned(4); struct mcx_cmd_init_hca_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_init_hca_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_teardown_hca_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[2]; #define MCX_CMD_TEARDOWN_HCA_GRACEFUL 0x0 #define MCX_CMD_TEARDOWN_HCA_PANIC 0x1 uint16_t cmd_profile; uint8_t cmd_reserved2[4]; } __packed __aligned(4); struct mcx_cmd_teardown_hca_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_access_reg_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[2]; uint16_t cmd_register_id; uint32_t cmd_argument; } __packed __aligned(4); struct mcx_cmd_access_reg_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_reg_pmtu { uint8_t rp_reserved1; uint8_t rp_local_port; uint8_t rp_reserved2[2]; uint16_t rp_max_mtu; uint8_t rp_reserved3[2]; uint16_t rp_admin_mtu; uint8_t rp_reserved4[2]; uint16_t rp_oper_mtu; uint8_t rp_reserved5[2]; } __packed __aligned(4); struct mcx_reg_ptys { uint8_t rp_reserved1; uint8_t rp_local_port; uint8_t rp_reserved2; uint8_t rp_proto_mask; #define MCX_REG_PTYS_PROTO_MASK_ETH (1 << 2) uint8_t rp_reserved3[8]; uint32_t rp_eth_proto_cap; uint8_t rp_reserved4[8]; uint32_t rp_eth_proto_admin; uint8_t rp_reserved5[8]; uint32_t rp_eth_proto_oper; uint8_t rp_reserved6[24]; } __packed __aligned(4); struct mcx_reg_paos { uint8_t rp_reserved1; uint8_t rp_local_port; uint8_t rp_admin_status; #define MCX_REG_PAOS_ADMIN_STATUS_UP 1 #define MCX_REG_PAOS_ADMIN_STATUS_DOWN 2 #define MCX_REG_PAOS_ADMIN_STATUS_UP_ONCE 3 #define MCX_REG_PAOS_ADMIN_STATUS_DISABLED 4 uint8_t rp_oper_status; #define MCX_REG_PAOS_OPER_STATUS_UP 1 #define MCX_REG_PAOS_OPER_STATUS_DOWN 2 #define MCX_REG_PAOS_OPER_STATUS_FAILED 4 uint8_t rp_admin_state_update; #define MCX_REG_PAOS_ADMIN_STATE_UPDATE_EN (1 << 7) uint8_t rp_reserved2[11]; } __packed __aligned(4); struct mcx_reg_pfcc { uint8_t rp_reserved1; uint8_t rp_local_port; uint8_t rp_reserved2[3]; uint8_t rp_prio_mask_tx; uint8_t rp_reserved3; uint8_t rp_prio_mask_rx; uint8_t rp_pptx_aptx; uint8_t rp_pfctx; uint8_t rp_fctx_dis; uint8_t rp_reserved4; uint8_t rp_pprx_aprx; uint8_t rp_pfcrx; uint8_t rp_reserved5[2]; uint16_t rp_dev_stall_min; uint16_t rp_dev_stall_crit; uint8_t rp_reserved6[12]; } __packed __aligned(4); #define MCX_PMLP_MODULE_NUM_MASK 0xff struct mcx_reg_pmlp { uint8_t rp_rxtx; uint8_t rp_local_port; uint8_t rp_reserved0; uint8_t rp_width; uint32_t rp_lane0_mapping; uint32_t rp_lane1_mapping; uint32_t rp_lane2_mapping; uint32_t rp_lane3_mapping; uint8_t rp_reserved1[44]; } __packed __aligned(4); struct mcx_reg_ppcnt { uint8_t ppcnt_swid; uint8_t ppcnt_local_port; uint8_t ppcnt_pnat; uint8_t ppcnt_grp; #define MCX_REG_PPCNT_GRP_IEEE8023 0x00 #define MCX_REG_PPCNT_GRP_RFC2863 0x01 #define MCX_REG_PPCNT_GRP_RFC2819 0x02 #define MCX_REG_PPCNT_GRP_RFC3635 0x03 #define MCX_REG_PPCNT_GRP_PER_PRIO 0x10 #define MCX_REG_PPCNT_GRP_PER_TC 0x11 #define MCX_REG_PPCNT_GRP_PER_RX_BUFFER 0x11 uint8_t ppcnt_clr; uint8_t ppcnt_reserved1[2]; uint8_t ppcnt_prio_tc; #define MCX_REG_PPCNT_CLR (1 << 7) uint8_t ppcnt_counter_set[248]; } __packed __aligned(8); CTASSERT(sizeof(struct mcx_reg_ppcnt) == 256); CTASSERT((offsetof(struct mcx_reg_ppcnt, ppcnt_counter_set) % sizeof(uint64_t)) == 0); enum mcx_ppcnt_ieee8023 { frames_transmitted_ok, frames_received_ok, frame_check_sequence_errors, alignment_errors, octets_transmitted_ok, octets_received_ok, multicast_frames_xmitted_ok, broadcast_frames_xmitted_ok, multicast_frames_received_ok, broadcast_frames_received_ok, in_range_length_errors, out_of_range_length_field, frame_too_long_errors, symbol_error_during_carrier, mac_control_frames_transmitted, mac_control_frames_received, unsupported_opcodes_received, pause_mac_ctrl_frames_received, pause_mac_ctrl_frames_transmitted, mcx_ppcnt_ieee8023_count }; CTASSERT(mcx_ppcnt_ieee8023_count * sizeof(uint64_t) == 0x98); enum mcx_ppcnt_rfc2863 { in_octets, in_ucast_pkts, in_discards, in_errors, in_unknown_protos, out_octets, out_ucast_pkts, out_discards, out_errors, in_multicast_pkts, in_broadcast_pkts, out_multicast_pkts, out_broadcast_pkts, mcx_ppcnt_rfc2863_count }; CTASSERT(mcx_ppcnt_rfc2863_count * sizeof(uint64_t) == 0x68); enum mcx_ppcnt_rfc2819 { drop_events, octets, pkts, broadcast_pkts, multicast_pkts, crc_align_errors, undersize_pkts, oversize_pkts, fragments, jabbers, collisions, pkts64octets, pkts65to127octets, pkts128to255octets, pkts256to511octets, pkts512to1023octets, pkts1024to1518octets, pkts1519to2047octets, pkts2048to4095octets, pkts4096to8191octets, pkts8192to10239octets, mcx_ppcnt_rfc2819_count }; CTASSERT((mcx_ppcnt_rfc2819_count * sizeof(uint64_t)) == 0xa8); enum mcx_ppcnt_rfc3635 { dot3stats_alignment_errors, dot3stats_fcs_errors, dot3stats_single_collision_frames, dot3stats_multiple_collision_frames, dot3stats_sqe_test_errors, dot3stats_deferred_transmissions, dot3stats_late_collisions, dot3stats_excessive_collisions, dot3stats_internal_mac_transmit_errors, dot3stats_carrier_sense_errors, dot3stats_frame_too_longs, dot3stats_internal_mac_receive_errors, dot3stats_symbol_errors, dot3control_in_unknown_opcodes, dot3in_pause_frames, dot3out_pause_frames, mcx_ppcnt_rfc3635_count }; CTASSERT((mcx_ppcnt_rfc3635_count * sizeof(uint64_t)) == 0x80); struct mcx_reg_mcam { uint8_t _reserved1[1]; uint8_t mcam_feature_group; uint8_t _reserved2[1]; uint8_t mcam_access_reg_group; uint8_t _reserved3[4]; uint8_t mcam_access_reg_cap_mask[16]; uint8_t _reserved4[16]; uint8_t mcam_feature_cap_mask[16]; uint8_t _reserved5[16]; } __packed __aligned(4); #define MCX_BITFIELD_BIT(bf, b) (bf[(sizeof bf - 1) - (b / 8)] & (b % 8)) #define MCX_MCAM_FEATURE_CAP_SENSOR_MAP 6 struct mcx_reg_mtcap { uint8_t _reserved1[3]; uint8_t mtcap_sensor_count; uint8_t _reserved2[4]; uint64_t mtcap_sensor_map; }; struct mcx_reg_mtmp { uint8_t _reserved1[2]; uint16_t mtmp_sensor_index; uint8_t _reserved2[2]; uint16_t mtmp_temperature; uint16_t mtmp_mte_mtr; #define MCX_REG_MTMP_MTE (1 << 15) #define MCX_REG_MTMP_MTR (1 << 14) uint16_t mtmp_max_temperature; uint16_t mtmp_tee; #define MCX_REG_MTMP_TEE_NOPE (0 << 14) #define MCX_REG_MTMP_TEE_GENERATE (1 << 14) #define MCX_REG_MTMP_TEE_GENERATE_ONE (2 << 14) uint16_t mtmp_temperature_threshold_hi; uint8_t _reserved3[2]; uint16_t mtmp_temperature_threshold_lo; uint8_t _reserved4[4]; uint8_t mtmp_sensor_name[8]; }; CTASSERT(sizeof(struct mcx_reg_mtmp) == 0x20); CTASSERT(offsetof(struct mcx_reg_mtmp, mtmp_sensor_name) == 0x18); #define MCX_MCIA_EEPROM_BYTES 32 struct mcx_reg_mcia { uint8_t rm_l; uint8_t rm_module; uint8_t rm_reserved0; uint8_t rm_status; uint8_t rm_i2c_addr; uint8_t rm_page_num; uint16_t rm_dev_addr; uint16_t rm_reserved1; uint16_t rm_size; uint32_t rm_reserved2; uint8_t rm_data[48]; } __packed __aligned(4); struct mcx_cmd_query_issi_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_query_issi_il_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[2]; uint16_t cmd_current_issi; uint8_t cmd_reserved2[4]; } __packed __aligned(4); CTASSERT(sizeof(struct mcx_cmd_query_issi_il_out) == MCX_CMDQ_INLINE_DATASIZE); struct mcx_cmd_query_issi_mb_out { uint8_t cmd_reserved2[16]; uint8_t cmd_supported_issi[80]; /* very big endian */ } __packed __aligned(4); CTASSERT(sizeof(struct mcx_cmd_query_issi_mb_out) <= MCX_CMDQ_MAILBOX_DATASIZE); struct mcx_cmd_set_issi_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[2]; uint16_t cmd_current_issi; uint8_t cmd_reserved2[4]; } __packed __aligned(4); CTASSERT(sizeof(struct mcx_cmd_set_issi_in) <= MCX_CMDQ_INLINE_DATASIZE); struct mcx_cmd_set_issi_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[8]; } __packed __aligned(4); CTASSERT(sizeof(struct mcx_cmd_set_issi_out) <= MCX_CMDQ_INLINE_DATASIZE); struct mcx_cmd_query_pages_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; #define MCX_CMD_QUERY_PAGES_BOOT 0x01 #define MCX_CMD_QUERY_PAGES_INIT 0x02 #define MCX_CMD_QUERY_PAGES_REGULAR 0x03 uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_query_pages_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[2]; uint16_t cmd_func_id; int32_t cmd_num_pages; } __packed __aligned(4); struct mcx_cmd_manage_pages_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; #define MCX_CMD_MANAGE_PAGES_ALLOC_FAIL \ 0x00 #define MCX_CMD_MANAGE_PAGES_ALLOC_SUCCESS \ 0x01 #define MCX_CMD_MANAGE_PAGES_HCA_RETURN_PAGES \ 0x02 uint8_t cmd_reserved1[2]; uint16_t cmd_func_id; uint32_t cmd_input_num_entries; } __packed __aligned(4); CTASSERT(sizeof(struct mcx_cmd_manage_pages_in) == MCX_CMDQ_INLINE_DATASIZE); struct mcx_cmd_manage_pages_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint32_t cmd_output_num_entries; uint8_t cmd_reserved1[4]; } __packed __aligned(4); CTASSERT(sizeof(struct mcx_cmd_manage_pages_out) == MCX_CMDQ_INLINE_DATASIZE); struct mcx_cmd_query_hca_cap_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; #define MCX_CMD_QUERY_HCA_CAP_MAX (0x0 << 0) #define MCX_CMD_QUERY_HCA_CAP_CURRENT (0x1 << 0) #define MCX_CMD_QUERY_HCA_CAP_DEVICE (0x0 << 1) #define MCX_CMD_QUERY_HCA_CAP_OFFLOAD (0x1 << 1) #define MCX_CMD_QUERY_HCA_CAP_FLOW (0x7 << 1) uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_query_hca_cap_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[8]; } __packed __aligned(4); #define MCX_HCA_CAP_LEN 0x1000 #define MCX_HCA_CAP_NMAILBOXES \ (MCX_HCA_CAP_LEN / MCX_CMDQ_MAILBOX_DATASIZE) #if __GNUC_PREREQ__(4, 3) #define __counter__ __COUNTER__ #else #define __counter__ __LINE__ #endif #define __token(_tok, _num) _tok##_num #define _token(_tok, _num) __token(_tok, _num) #define __reserved__ _token(__reserved, __counter__) struct mcx_cap_device { uint8_t reserved0[16]; uint8_t log_max_srq_sz; uint8_t log_max_qp_sz; uint8_t __reserved__[1]; uint8_t log_max_qp; /* 5 bits */ #define MCX_CAP_DEVICE_LOG_MAX_QP 0x1f uint8_t __reserved__[1]; uint8_t log_max_srq; /* 5 bits */ #define MCX_CAP_DEVICE_LOG_MAX_SRQ 0x1f uint8_t __reserved__[2]; uint8_t __reserved__[1]; uint8_t log_max_cq_sz; uint8_t __reserved__[1]; uint8_t log_max_cq; /* 5 bits */ #define MCX_CAP_DEVICE_LOG_MAX_CQ 0x1f uint8_t log_max_eq_sz; uint8_t log_max_mkey; /* 6 bits */ #define MCX_CAP_DEVICE_LOG_MAX_MKEY 0x3f uint8_t __reserved__[1]; uint8_t log_max_eq; /* 4 bits */ #define MCX_CAP_DEVICE_LOG_MAX_EQ 0x0f uint8_t max_indirection; uint8_t log_max_mrw_sz; /* 7 bits */ #define MCX_CAP_DEVICE_LOG_MAX_MRW_SZ 0x7f uint8_t teardown_log_max_msf_list_size; #define MCX_CAP_DEVICE_FORCE_TEARDOWN 0x80 #define MCX_CAP_DEVICE_LOG_MAX_MSF_LIST_SIZE \ 0x3f uint8_t log_max_klm_list_size; /* 6 bits */ #define MCX_CAP_DEVICE_LOG_MAX_KLM_LIST_SIZE \ 0x3f uint8_t __reserved__[1]; uint8_t log_max_ra_req_dc; /* 6 bits */ #define MCX_CAP_DEVICE_LOG_MAX_REQ_DC 0x3f uint8_t __reserved__[1]; uint8_t log_max_ra_res_dc; /* 6 bits */ #define MCX_CAP_DEVICE_LOG_MAX_RA_RES_DC \ 0x3f uint8_t __reserved__[1]; uint8_t log_max_ra_req_qp; /* 6 bits */ #define MCX_CAP_DEVICE_LOG_MAX_RA_REQ_QP \ 0x3f uint8_t __reserved__[1]; uint8_t log_max_ra_res_qp; /* 6 bits */ #define MCX_CAP_DEVICE_LOG_MAX_RA_RES_QP \ 0x3f uint8_t flags1; #define MCX_CAP_DEVICE_END_PAD 0x80 #define MCX_CAP_DEVICE_CC_QUERY_ALLOWED 0x40 #define MCX_CAP_DEVICE_CC_MODIFY_ALLOWED \ 0x20 #define MCX_CAP_DEVICE_START_PAD 0x10 #define MCX_CAP_DEVICE_128BYTE_CACHELINE \ 0x08 uint8_t __reserved__[1]; uint16_t gid_table_size; uint16_t flags2; #define MCX_CAP_DEVICE_OUT_OF_SEQ_CNT 0x8000 #define MCX_CAP_DEVICE_VPORT_COUNTERS 0x4000 #define MCX_CAP_DEVICE_RETRANSMISSION_Q_COUNTERS \ 0x2000 #define MCX_CAP_DEVICE_DEBUG 0x1000 #define MCX_CAP_DEVICE_MODIFY_RQ_COUNTERS_SET_ID \ 0x8000 #define MCX_CAP_DEVICE_RQ_DELAY_DROP 0x4000 #define MCX_CAP_DEVICe_MAX_QP_CNT_MASK 0x03ff uint16_t pkey_table_size; uint8_t flags3; #define MCX_CAP_DEVICE_VPORT_GROUP_MANAGER \ 0x80 #define MCX_CAP_DEVICE_VHCA_GROUP_MANAGER \ 0x40 #define MCX_CAP_DEVICE_IB_VIRTUAL 0x20 #define MCX_CAP_DEVICE_ETH_VIRTUAL 0x10 #define MCX_CAP_DEVICE_ETS 0x04 #define MCX_CAP_DEVICE_NIC_FLOW_TABLE 0x02 #define MCX_CAP_DEVICE_ESWITCH_FLOW_TABLE \ 0x01 uint8_t local_ca_ack_delay; /* 5 bits */ #define MCX_CAP_DEVICE_LOCAL_CA_ACK_DELAY \ 0x1f #define MCX_CAP_DEVICE_MCAM_REG 0x40 uint8_t port_type; #define MCX_CAP_DEVICE_PORT_MODULE_EVENT \ 0x80 #define MCX_CAP_DEVICE_PORT_TYPE 0x03 #define MCX_CAP_DEVICE_PORT_TYPE_ETH 0x01 uint8_t num_ports; uint8_t snapshot_log_max_msg; #define MCX_CAP_DEVICE_SNAPSHOT 0x80 #define MCX_CAP_DEVICE_LOG_MAX_MSG 0x1f uint8_t max_tc; /* 4 bits */ #define MCX_CAP_DEVICE_MAX_TC 0x0f uint8_t flags4; #define MCX_CAP_DEVICE_TEMP_WARN_EVENT 0x80 #define MCX_CAP_DEVICE_DCBX 0x40 #define MCX_CAP_DEVICE_ROL_S 0x02 #define MCX_CAP_DEVICE_ROL_G 0x01 uint8_t wol; #define MCX_CAP_DEVICE_WOL_S 0x40 #define MCX_CAP_DEVICE_WOL_G 0x20 #define MCX_CAP_DEVICE_WOL_A 0x10 #define MCX_CAP_DEVICE_WOL_B 0x08 #define MCX_CAP_DEVICE_WOL_M 0x04 #define MCX_CAP_DEVICE_WOL_U 0x02 #define MCX_CAP_DEVICE_WOL_P 0x01 uint16_t stat_rate_support; uint8_t __reserved__[1]; uint8_t cqe_version; /* 4 bits */ #define MCX_CAP_DEVICE_CQE_VERSION 0x0f uint32_t flags5; #define MCX_CAP_DEVICE_COMPACT_ADDRESS_VECTOR \ 0x80000000 #define MCX_CAP_DEVICE_STRIDING_RQ 0x40000000 #define MCX_CAP_DEVICE_IPOIP_ENHANCED_OFFLOADS \ 0x10000000 #define MCX_CAP_DEVICE_IPOIP_IPOIP_OFFLOADS \ 0x08000000 #define MCX_CAP_DEVICE_DC_CONNECT_CP 0x00040000 #define MCX_CAP_DEVICE_DC_CNAK_DRACE 0x00020000 #define MCX_CAP_DEVICE_DRAIN_SIGERR 0x00010000 #define MCX_CAP_DEVICE_DRAIN_SIGERR 0x00010000 #define MCX_CAP_DEVICE_CMDIF_CHECKSUM 0x0000c000 #define MCX_CAP_DEVICE_SIGERR_QCE 0x00002000 #define MCX_CAP_DEVICE_WQ_SIGNATURE 0x00000800 #define MCX_CAP_DEVICE_SCTR_DATA_CQE 0x00000400 #define MCX_CAP_DEVICE_SHO 0x00000100 #define MCX_CAP_DEVICE_TPH 0x00000080 #define MCX_CAP_DEVICE_RF 0x00000040 #define MCX_CAP_DEVICE_DCT 0x00000020 #define MCX_CAP_DEVICE_QOS 0x00000010 #define MCX_CAP_DEVICe_ETH_NET_OFFLOADS 0x00000008 #define MCX_CAP_DEVICE_ROCE 0x00000004 #define MCX_CAP_DEVICE_ATOMIC 0x00000002 uint32_t flags6; #define MCX_CAP_DEVICE_CQ_OI 0x80000000 #define MCX_CAP_DEVICE_CQ_RESIZE 0x40000000 #define MCX_CAP_DEVICE_CQ_MODERATION 0x20000000 #define MCX_CAP_DEVICE_CQ_PERIOD_MODE_MODIFY \ 0x10000000 #define MCX_CAP_DEVICE_CQ_INVALIDATE 0x08000000 #define MCX_CAP_DEVICE_RESERVED_AT_255 0x04000000 #define MCX_CAP_DEVICE_CQ_EQ_REMAP 0x02000000 #define MCX_CAP_DEVICE_PG 0x01000000 #define MCX_CAP_DEVICE_BLOCK_LB_MC 0x00800000 #define MCX_CAP_DEVICE_EXPONENTIAL_BACKOFF \ 0x00400000 #define MCX_CAP_DEVICE_SCQE_BREAK_MODERATION \ 0x00200000 #define MCX_CAP_DEVICE_CQ_PERIOD_START_FROM_CQE \ 0x00100000 #define MCX_CAP_DEVICE_CD 0x00080000 #define MCX_CAP_DEVICE_ATM 0x00040000 #define MCX_CAP_DEVICE_APM 0x00020000 #define MCX_CAP_DEVICE_IMAICL 0x00010000 #define MCX_CAP_DEVICE_QKV 0x00000200 #define MCX_CAP_DEVICE_PKV 0x00000100 #define MCX_CAP_DEVICE_SET_DETH_SQPN 0x00000080 #define MCX_CAP_DEVICE_XRC 0x00000008 #define MCX_CAP_DEVICE_UD 0x00000004 #define MCX_CAP_DEVICE_UC 0x00000002 #define MCX_CAP_DEVICE_RC 0x00000001 uint8_t uar_flags; #define MCX_CAP_DEVICE_UAR_4K 0x80 uint8_t uar_sz; /* 6 bits */ #define MCX_CAP_DEVICE_UAR_SZ 0x3f uint8_t __reserved__[1]; uint8_t log_pg_sz; uint8_t flags7; #define MCX_CAP_DEVICE_BF 0x80 #define MCX_CAP_DEVICE_DRIVER_VERSION 0x40 #define MCX_CAP_DEVICE_PAD_TX_ETH_PACKET \ 0x20 uint8_t log_bf_reg_size; /* 5 bits */ #define MCX_CAP_DEVICE_LOG_BF_REG_SIZE 0x1f uint8_t __reserved__[2]; uint16_t num_of_diagnostic_counters; uint16_t max_wqe_sz_sq; uint8_t __reserved__[2]; uint16_t max_wqe_sz_rq; uint8_t __reserved__[2]; uint16_t max_wqe_sz_sq_dc; uint32_t max_qp_mcg; /* 25 bits */ #define MCX_CAP_DEVICE_MAX_QP_MCG 0x1ffffff uint8_t __reserved__[3]; uint8_t log_max_mcq; uint8_t log_max_transport_domain; /* 5 bits */ #define MCX_CAP_DEVICE_LOG_MAX_TRANSORT_DOMAIN \ 0x1f uint8_t log_max_pd; /* 5 bits */ #define MCX_CAP_DEVICE_LOG_MAX_PD 0x1f uint8_t __reserved__[1]; uint8_t log_max_xrcd; /* 5 bits */ #define MCX_CAP_DEVICE_LOG_MAX_XRCD 0x1f uint8_t __reserved__[2]; uint16_t max_flow_counter; uint8_t log_max_rq; /* 5 bits */ #define MCX_CAP_DEVICE_LOG_MAX_RQ 0x1f uint8_t log_max_sq; /* 5 bits */ #define MCX_CAP_DEVICE_LOG_MAX_SQ 0x1f uint8_t log_max_tir; /* 5 bits */ #define MCX_CAP_DEVICE_LOG_MAX_TIR 0x1f uint8_t log_max_tis; /* 5 bits */ #define MCX_CAP_DEVICE_LOG_MAX_TIS 0x1f uint8_t flags8; #define MCX_CAP_DEVICE_BASIC_CYCLIC_RCV_WQE \ 0x80 #define MCX_CAP_DEVICE_LOG_MAX_RMP 0x1f uint8_t log_max_rqt; /* 5 bits */ #define MCX_CAP_DEVICE_LOG_MAX_RQT 0x1f uint8_t log_max_rqt_size; /* 5 bits */ #define MCX_CAP_DEVICE_LOG_MAX_RQT_SIZE 0x1f uint8_t log_max_tis_per_sq; /* 5 bits */ #define MCX_CAP_DEVICE_LOG_MAX_TIS_PER_SQ \ 0x1f uint8_t flags9; #define MXC_CAP_DEVICE_EXT_STRIDE_NUM_RANGES \ 0x80 #define MXC_CAP_DEVICE_LOG_MAX_STRIDE_SZ_RQ \ 0x1f uint8_t log_min_stride_sz_rq; /* 5 bits */ #define MXC_CAP_DEVICE_LOG_MIN_STRIDE_SZ_RQ \ 0x1f uint8_t log_max_stride_sz_sq; /* 5 bits */ #define MXC_CAP_DEVICE_LOG_MAX_STRIDE_SZ_SQ \ 0x1f uint8_t log_min_stride_sz_sq; /* 5 bits */ #define MXC_CAP_DEVICE_LOG_MIN_STRIDE_SZ_SQ \ 0x1f uint8_t log_max_hairpin_queues; #define MXC_CAP_DEVICE_HAIRPIN 0x80 #define MXC_CAP_DEVICE_LOG_MAX_HAIRPIN_QUEUES \ 0x1f uint8_t log_min_hairpin_queues; #define MXC_CAP_DEVICE_LOG_MIN_HAIRPIN_QUEUES \ 0x1f uint8_t log_max_hairpin_num_packets; #define MXC_CAP_DEVICE_LOG_MAX_HAIRPIN_NUM_PACKETS \ 0x1f uint8_t log_max_mq_sz; #define MXC_CAP_DEVICE_LOG_MAX_WQ_SZ \ 0x1f uint8_t log_min_hairpin_wq_data_sz; #define MXC_CAP_DEVICE_NIC_VPORT_CHANGE_EVENT \ 0x80 #define MXC_CAP_DEVICE_DISABLE_LOCAL_LB_UC \ 0x40 #define MXC_CAP_DEVICE_DISABLE_LOCAL_LB_MC \ 0x20 #define MCX_CAP_DEVICE_LOG_MIN_HAIRPIN_WQ_DATA_SZ \ 0x1f uint8_t log_max_vlan_list; #define MXC_CAP_DEVICE_SYSTEM_IMAGE_GUID_MODIFIABLE \ 0x80 #define MXC_CAP_DEVICE_LOG_MAX_VLAN_LIST \ 0x1f uint8_t log_max_current_mc_list; #define MXC_CAP_DEVICE_LOG_MAX_CURRENT_MC_LIST \ 0x1f uint8_t log_max_current_uc_list; #define MXC_CAP_DEVICE_LOG_MAX_CURRENT_UC_LIST \ 0x1f uint8_t __reserved__[4]; uint32_t create_qp_start_hint; /* 24 bits */ uint8_t log_max_uctx; /* 5 bits */ #define MXC_CAP_DEVICE_LOG_MAX_UCTX 0x1f uint8_t log_max_umem; /* 5 bits */ #define MXC_CAP_DEVICE_LOG_MAX_UMEM 0x1f uint16_t max_num_eqs; uint8_t log_max_l2_table; /* 5 bits */ #define MXC_CAP_DEVICE_LOG_MAX_L2_TABLE 0x1f uint8_t __reserved__[1]; uint16_t log_uar_page_sz; uint8_t __reserved__[8]; uint32_t device_frequency_mhz; uint32_t device_frequency_khz; } __packed __aligned(8); CTASSERT(offsetof(struct mcx_cap_device, max_indirection) == 0x20); CTASSERT(offsetof(struct mcx_cap_device, flags1) == 0x2c); CTASSERT(offsetof(struct mcx_cap_device, flags2) == 0x30); CTASSERT(offsetof(struct mcx_cap_device, snapshot_log_max_msg) == 0x38); CTASSERT(offsetof(struct mcx_cap_device, flags5) == 0x40); CTASSERT(offsetof(struct mcx_cap_device, flags7) == 0x4c); CTASSERT(offsetof(struct mcx_cap_device, device_frequency_mhz) == 0x98); CTASSERT(offsetof(struct mcx_cap_device, device_frequency_khz) == 0x9c); CTASSERT(sizeof(struct mcx_cap_device) <= MCX_CMDQ_MAILBOX_DATASIZE); struct mcx_cmd_set_driver_version_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_set_driver_version_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_set_driver_version { uint8_t cmd_driver_version[64]; } __packed __aligned(8); struct mcx_cmd_modify_nic_vport_context_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[4]; uint32_t cmd_field_select; #define MCX_CMD_MODIFY_NIC_VPORT_CONTEXT_FIELD_ADDR 0x04 #define MCX_CMD_MODIFY_NIC_VPORT_CONTEXT_FIELD_PROMISC 0x10 #define MCX_CMD_MODIFY_NIC_VPORT_CONTEXT_FIELD_MTU 0x40 } __packed __aligned(4); struct mcx_cmd_modify_nic_vport_context_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_query_nic_vport_context_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[4]; uint8_t cmd_allowed_list_type; uint8_t cmd_reserved2[3]; } __packed __aligned(4); struct mcx_cmd_query_nic_vport_context_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_nic_vport_ctx { uint32_t vp_min_wqe_inline_mode; uint8_t vp_reserved0[32]; uint32_t vp_mtu; uint8_t vp_reserved1[200]; uint16_t vp_flags; #define MCX_NIC_VPORT_CTX_LIST_UC_MAC (0) #define MCX_NIC_VPORT_CTX_LIST_MC_MAC (1 << 24) #define MCX_NIC_VPORT_CTX_LIST_VLAN (2 << 24) #define MCX_NIC_VPORT_CTX_PROMISC_ALL (1 << 13) #define MCX_NIC_VPORT_CTX_PROMISC_MCAST (1 << 14) #define MCX_NIC_VPORT_CTX_PROMISC_UCAST (1 << 15) uint16_t vp_allowed_list_size; uint64_t vp_perm_addr; uint8_t vp_reserved2[4]; /* allowed list follows */ } __packed __aligned(4); struct mcx_counter { uint64_t packets; uint64_t octets; } __packed __aligned(4); struct mcx_nic_vport_counters { struct mcx_counter rx_err; struct mcx_counter tx_err; uint8_t reserved0[64]; /* 0x30 */ struct mcx_counter rx_bcast; struct mcx_counter tx_bcast; struct mcx_counter rx_ucast; struct mcx_counter tx_ucast; struct mcx_counter rx_mcast; struct mcx_counter tx_mcast; uint8_t reserved1[0x210 - 0xd0]; } __packed __aligned(4); struct mcx_cmd_query_vport_counters_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_query_vport_counters_mb_in { uint8_t cmd_reserved0[8]; uint8_t cmd_clear; uint8_t cmd_reserved1[7]; } __packed __aligned(4); struct mcx_cmd_query_vport_counters_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_query_flow_counter_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_query_flow_counter_mb_in { uint8_t cmd_reserved0[8]; uint8_t cmd_clear; uint8_t cmd_reserved1[5]; uint16_t cmd_flow_counter_id; } __packed __aligned(4); struct mcx_cmd_query_flow_counter_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_alloc_uar_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_alloc_uar_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint32_t cmd_uar; uint8_t cmd_reserved1[4]; } __packed __aligned(4); struct mcx_cmd_query_special_ctx_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_query_special_ctx_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[4]; uint32_t cmd_resd_lkey; } __packed __aligned(4); struct mcx_eq_ctx { uint32_t eq_status; #define MCX_EQ_CTX_STATE_SHIFT 8 #define MCX_EQ_CTX_STATE_MASK (0xf << MCX_EQ_CTX_STATE_SHIFT) #define MCX_EQ_CTX_STATE_ARMED 0x9 #define MCX_EQ_CTX_STATE_FIRED 0xa #define MCX_EQ_CTX_OI_SHIFT 17 #define MCX_EQ_CTX_OI (1 << MCX_EQ_CTX_OI_SHIFT) #define MCX_EQ_CTX_EC_SHIFT 18 #define MCX_EQ_CTX_EC (1 << MCX_EQ_CTX_EC_SHIFT) #define MCX_EQ_CTX_STATUS_SHIFT 28 #define MCX_EQ_CTX_STATUS_MASK (0xf << MCX_EQ_CTX_STATUS_SHIFT) #define MCX_EQ_CTX_STATUS_OK 0x0 #define MCX_EQ_CTX_STATUS_EQ_WRITE_FAILURE 0xa uint32_t eq_reserved1; uint32_t eq_page_offset; #define MCX_EQ_CTX_PAGE_OFFSET_SHIFT 5 uint32_t eq_uar_size; #define MCX_EQ_CTX_UAR_PAGE_MASK 0xffffff #define MCX_EQ_CTX_LOG_EQ_SIZE_SHIFT 24 uint32_t eq_reserved2; uint8_t eq_reserved3[3]; uint8_t eq_intr; uint32_t eq_log_page_size; #define MCX_EQ_CTX_LOG_PAGE_SIZE_SHIFT 24 uint32_t eq_reserved4[3]; uint32_t eq_consumer_counter; uint32_t eq_producer_counter; #define MCX_EQ_CTX_COUNTER_MASK 0xffffff uint32_t eq_reserved5[4]; } __packed __aligned(4); CTASSERT(sizeof(struct mcx_eq_ctx) == 64); struct mcx_cmd_create_eq_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_create_eq_mb_in { struct mcx_eq_ctx cmd_eq_ctx; uint8_t cmd_reserved0[8]; uint64_t cmd_event_bitmask; #define MCX_EVENT_TYPE_COMPLETION 0x00 #define MCX_EVENT_TYPE_CQ_ERROR 0x04 #define MCX_EVENT_TYPE_INTERNAL_ERROR 0x08 #define MCX_EVENT_TYPE_PORT_CHANGE 0x09 #define MCX_EVENT_TYPE_CMD_COMPLETION 0x0a #define MCX_EVENT_TYPE_PAGE_REQUEST 0x0b #define MCX_EVENT_TYPE_LAST_WQE 0x13 uint8_t cmd_reserved1[176]; } __packed __aligned(4); struct mcx_cmd_create_eq_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint32_t cmd_eqn; uint8_t cmd_reserved1[4]; } __packed __aligned(4); struct mcx_cmd_query_eq_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint32_t cmd_eqn; uint8_t cmd_reserved1[4]; } __packed __aligned(4); struct mcx_cmd_query_eq_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_eq_entry { uint8_t eq_reserved1; uint8_t eq_event_type; uint8_t eq_reserved2; uint8_t eq_event_sub_type; uint8_t eq_reserved3[28]; uint32_t eq_event_data[7]; uint8_t eq_reserved4[2]; uint8_t eq_signature; uint8_t eq_owner; #define MCX_EQ_ENTRY_OWNER_INIT 1 } __packed __aligned(4); CTASSERT(sizeof(struct mcx_eq_entry) == 64); struct mcx_cmd_alloc_pd_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_alloc_pd_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint32_t cmd_pd; uint8_t cmd_reserved1[4]; } __packed __aligned(4); struct mcx_cmd_alloc_td_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_alloc_td_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint32_t cmd_tdomain; uint8_t cmd_reserved1[4]; } __packed __aligned(4); struct mcx_cmd_create_tir_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_create_tir_mb_in { uint8_t cmd_reserved0[20]; uint32_t cmd_disp_type; #define MCX_TIR_CTX_DISP_TYPE_DIRECT 0 #define MCX_TIR_CTX_DISP_TYPE_INDIRECT 1 #define MCX_TIR_CTX_DISP_TYPE_SHIFT 28 uint8_t cmd_reserved1[8]; uint32_t cmd_lro; uint8_t cmd_reserved2[8]; uint32_t cmd_inline_rqn; uint32_t cmd_indir_table; uint32_t cmd_tdomain; #define MCX_TIR_CTX_HASH_TOEPLITZ 2 #define MCX_TIR_CTX_HASH_SHIFT 28 uint8_t cmd_rx_hash_key[40]; uint32_t cmd_rx_hash_sel_outer; #define MCX_TIR_CTX_HASH_SEL_SRC_IP (1 << 0) #define MCX_TIR_CTX_HASH_SEL_DST_IP (1 << 1) #define MCX_TIR_CTX_HASH_SEL_SPORT (1 << 2) #define MCX_TIR_CTX_HASH_SEL_DPORT (1 << 3) #define MCX_TIR_CTX_HASH_SEL_IPV4 (0 << 31) #define MCX_TIR_CTX_HASH_SEL_IPV6 (1 << 31) #define MCX_TIR_CTX_HASH_SEL_TCP (0 << 30) #define MCX_TIR_CTX_HASH_SEL_UDP (1 << 30) uint32_t cmd_rx_hash_sel_inner; uint8_t cmd_reserved3[152]; } __packed __aligned(4); struct mcx_cmd_create_tir_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint32_t cmd_tirn; uint8_t cmd_reserved1[4]; } __packed __aligned(4); struct mcx_cmd_destroy_tir_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint32_t cmd_tirn; uint8_t cmd_reserved1[4]; } __packed __aligned(4); struct mcx_cmd_destroy_tir_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_create_tis_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_create_tis_mb_in { uint8_t cmd_reserved[16]; uint32_t cmd_prio; uint8_t cmd_reserved1[32]; uint32_t cmd_tdomain; uint8_t cmd_reserved2[120]; } __packed __aligned(4); struct mcx_cmd_create_tis_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint32_t cmd_tisn; uint8_t cmd_reserved1[4]; } __packed __aligned(4); struct mcx_cmd_destroy_tis_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint32_t cmd_tisn; uint8_t cmd_reserved1[4]; } __packed __aligned(4); struct mcx_cmd_destroy_tis_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_create_rqt_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_rqt_ctx { uint8_t cmd_reserved0[20]; uint16_t cmd_reserved1; uint16_t cmd_rqt_max_size; uint16_t cmd_reserved2; uint16_t cmd_rqt_actual_size; uint8_t cmd_reserved3[212]; } __packed __aligned(4); struct mcx_cmd_create_rqt_mb_in { uint8_t cmd_reserved0[16]; struct mcx_rqt_ctx cmd_rqt; } __packed __aligned(4); struct mcx_cmd_create_rqt_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint32_t cmd_rqtn; uint8_t cmd_reserved1[4]; } __packed __aligned(4); struct mcx_cmd_destroy_rqt_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint32_t cmd_rqtn; uint8_t cmd_reserved1[4]; } __packed __aligned(4); struct mcx_cmd_destroy_rqt_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cq_ctx { uint32_t cq_status; #define MCX_CQ_CTX_STATUS_SHIFT 28 #define MCX_CQ_CTX_STATUS_MASK (0xf << MCX_CQ_CTX_STATUS_SHIFT) #define MCX_CQ_CTX_STATUS_OK 0x0 #define MCX_CQ_CTX_STATUS_OVERFLOW 0x9 #define MCX_CQ_CTX_STATUS_WRITE_FAIL 0xa #define MCX_CQ_CTX_STATE_SHIFT 8 #define MCX_CQ_CTX_STATE_MASK (0xf << MCX_CQ_CTX_STATE_SHIFT) #define MCX_CQ_CTX_STATE_SOLICITED 0x6 #define MCX_CQ_CTX_STATE_ARMED 0x9 #define MCX_CQ_CTX_STATE_FIRED 0xa uint32_t cq_reserved1; uint32_t cq_page_offset; uint32_t cq_uar_size; #define MCX_CQ_CTX_UAR_PAGE_MASK 0xffffff #define MCX_CQ_CTX_LOG_CQ_SIZE_SHIFT 24 uint32_t cq_period_max_count; #define MCX_CQ_CTX_PERIOD_SHIFT 16 uint32_t cq_eqn; uint32_t cq_log_page_size; #define MCX_CQ_CTX_LOG_PAGE_SIZE_SHIFT 24 uint32_t cq_reserved2; uint32_t cq_last_notified; uint32_t cq_last_solicit; uint32_t cq_consumer_counter; uint32_t cq_producer_counter; uint8_t cq_reserved3[8]; uint64_t cq_doorbell; } __packed __aligned(4); CTASSERT(sizeof(struct mcx_cq_ctx) == 64); struct mcx_cmd_create_cq_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_create_cq_mb_in { struct mcx_cq_ctx cmd_cq_ctx; uint8_t cmd_reserved1[192]; } __packed __aligned(4); struct mcx_cmd_create_cq_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint32_t cmd_cqn; uint8_t cmd_reserved1[4]; } __packed __aligned(4); struct mcx_cmd_destroy_cq_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint32_t cmd_cqn; uint8_t cmd_reserved1[4]; } __packed __aligned(4); struct mcx_cmd_destroy_cq_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_query_cq_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint32_t cmd_cqn; uint8_t cmd_reserved1[4]; } __packed __aligned(4); struct mcx_cmd_query_cq_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cq_entry { uint32_t __reserved__; uint32_t cq_lro; uint32_t cq_lro_ack_seq_num; uint32_t cq_rx_hash; uint8_t cq_rx_hash_type; uint8_t cq_ml_path; uint16_t __reserved__; uint32_t cq_checksum; uint32_t __reserved__; uint32_t cq_flags; #define MCX_CQ_ENTRY_FLAGS_L4_OK (1 << 26) #define MCX_CQ_ENTRY_FLAGS_L3_OK (1 << 25) #define MCX_CQ_ENTRY_FLAGS_L2_OK (1 << 24) #define MCX_CQ_ENTRY_FLAGS_CV (1 << 16) #define MCX_CQ_ENTRY_FLAGS_VLAN_MASK (0xffff) uint32_t cq_lro_srqn; uint32_t __reserved__[2]; uint32_t cq_byte_cnt; uint64_t cq_timestamp; uint8_t cq_rx_drops; uint8_t cq_flow_tag[3]; uint16_t cq_wqe_count; uint8_t cq_signature; uint8_t cq_opcode_owner; #define MCX_CQ_ENTRY_FLAG_OWNER (1 << 0) #define MCX_CQ_ENTRY_FLAG_SE (1 << 1) #define MCX_CQ_ENTRY_FORMAT_SHIFT 2 #define MCX_CQ_ENTRY_OPCODE_SHIFT 4 #define MCX_CQ_ENTRY_FORMAT_NO_INLINE 0 #define MCX_CQ_ENTRY_FORMAT_INLINE_32 1 #define MCX_CQ_ENTRY_FORMAT_INLINE_64 2 #define MCX_CQ_ENTRY_FORMAT_COMPRESSED 3 #define MCX_CQ_ENTRY_OPCODE_REQ 0 #define MCX_CQ_ENTRY_OPCODE_SEND 2 #define MCX_CQ_ENTRY_OPCODE_REQ_ERR 13 #define MCX_CQ_ENTRY_OPCODE_SEND_ERR 14 #define MCX_CQ_ENTRY_OPCODE_INVALID 15 } __packed __aligned(4); CTASSERT(sizeof(struct mcx_cq_entry) == 64); struct mcx_cq_doorbell { uint32_t db_update_ci; uint32_t db_arm_ci; #define MCX_CQ_DOORBELL_ARM_CMD_SN_SHIFT 28 #define MCX_CQ_DOORBELL_ARM_CMD (1 << 24) #define MCX_CQ_DOORBELL_ARM_CI_MASK (0xffffff) } __packed __aligned(8); struct mcx_wq_ctx { uint8_t wq_type; #define MCX_WQ_CTX_TYPE_CYCLIC (1 << 4) #define MCX_WQ_CTX_TYPE_SIGNATURE (1 << 3) uint8_t wq_reserved0[5]; uint16_t wq_lwm; uint32_t wq_pd; uint32_t wq_uar_page; uint64_t wq_doorbell; uint32_t wq_hw_counter; uint32_t wq_sw_counter; uint16_t wq_log_stride; uint8_t wq_log_page_sz; uint8_t wq_log_size; uint8_t wq_reserved1[156]; } __packed __aligned(4); CTASSERT(sizeof(struct mcx_wq_ctx) == 0xC0); struct mcx_sq_ctx { uint32_t sq_flags; #define MCX_SQ_CTX_RLKEY (1U << 31) #define MCX_SQ_CTX_FRE_SHIFT (1 << 29) #define MCX_SQ_CTX_FLUSH_IN_ERROR (1 << 28) #define MCX_SQ_CTX_MIN_WQE_INLINE_SHIFT 24 #define MCX_SQ_CTX_STATE_SHIFT 20 #define MCX_SQ_CTX_STATE_MASK (0xf << 20) #define MCX_SQ_CTX_STATE_RST 0 #define MCX_SQ_CTX_STATE_RDY 1 #define MCX_SQ_CTX_STATE_ERR 3 uint32_t sq_user_index; uint32_t sq_cqn; uint32_t sq_reserved1[5]; uint32_t sq_tis_lst_sz; #define MCX_SQ_CTX_TIS_LST_SZ_SHIFT 16 uint32_t sq_reserved2[2]; uint32_t sq_tis_num; struct mcx_wq_ctx sq_wq; } __packed __aligned(4); struct mcx_sq_entry_seg { uint32_t sqs_byte_count; uint32_t sqs_lkey; uint64_t sqs_addr; } __packed __aligned(4); struct mcx_sq_entry { /* control segment */ uint32_t sqe_opcode_index; #define MCX_SQE_WQE_INDEX_SHIFT 8 #define MCX_SQE_WQE_OPCODE_NOP 0x00 #define MCX_SQE_WQE_OPCODE_SEND 0x0a uint32_t sqe_ds_sq_num; #define MCX_SQE_SQ_NUM_SHIFT 8 uint32_t sqe_signature; #define MCX_SQE_SIGNATURE_SHIFT 24 #define MCX_SQE_SOLICITED_EVENT 0x02 #define MCX_SQE_CE_CQE_ON_ERR 0x00 #define MCX_SQE_CE_CQE_FIRST_ERR 0x04 #define MCX_SQE_CE_CQE_ALWAYS 0x08 #define MCX_SQE_CE_CQE_SOLICIT 0x0C #define MCX_SQE_FM_NO_FENCE 0x00 #define MCX_SQE_FM_SMALL_FENCE 0x40 uint32_t sqe_mkey; /* ethernet segment */ uint32_t sqe_reserved1; uint32_t sqe_mss_csum; #define MCX_SQE_L4_CSUM (1 << 31) #define MCX_SQE_L3_CSUM (1 << 30) uint32_t sqe_reserved2; uint16_t sqe_inline_header_size; uint16_t sqe_inline_headers[9]; /* data segment */ struct mcx_sq_entry_seg sqe_segs[1]; } __packed __aligned(64); CTASSERT(sizeof(struct mcx_sq_entry) == 64); struct mcx_cmd_create_sq_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_create_sq_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint32_t cmd_sqn; uint8_t cmd_reserved1[4]; } __packed __aligned(4); struct mcx_cmd_modify_sq_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint32_t cmd_sq_state; uint8_t cmd_reserved1[4]; } __packed __aligned(4); struct mcx_cmd_modify_sq_mb_in { uint32_t cmd_modify_hi; uint32_t cmd_modify_lo; uint8_t cmd_reserved0[8]; struct mcx_sq_ctx cmd_sq_ctx; } __packed __aligned(4); struct mcx_cmd_modify_sq_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_destroy_sq_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint32_t cmd_sqn; uint8_t cmd_reserved1[4]; } __packed __aligned(4); struct mcx_cmd_destroy_sq_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_rq_ctx { uint32_t rq_flags; #define MCX_RQ_CTX_RLKEY (1U << 31) #define MCX_RQ_CTX_VLAN_STRIP_DIS (1 << 28) #define MCX_RQ_CTX_MEM_RQ_TYPE_SHIFT 24 #define MCX_RQ_CTX_STATE_SHIFT 20 #define MCX_RQ_CTX_STATE_MASK (0xf << 20) #define MCX_RQ_CTX_STATE_RST 0 #define MCX_RQ_CTX_STATE_RDY 1 #define MCX_RQ_CTX_STATE_ERR 3 #define MCX_RQ_CTX_FLUSH_IN_ERROR (1 << 18) uint32_t rq_user_index; uint32_t rq_cqn; uint32_t rq_reserved1; uint32_t rq_rmpn; uint32_t rq_reserved2[7]; struct mcx_wq_ctx rq_wq; } __packed __aligned(4); struct mcx_rq_entry { uint32_t rqe_byte_count; uint32_t rqe_lkey; uint64_t rqe_addr; } __packed __aligned(16); struct mcx_cmd_create_rq_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_create_rq_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint32_t cmd_rqn; uint8_t cmd_reserved1[4]; } __packed __aligned(4); struct mcx_cmd_modify_rq_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint32_t cmd_rq_state; uint8_t cmd_reserved1[4]; } __packed __aligned(4); struct mcx_cmd_modify_rq_mb_in { uint32_t cmd_modify_hi; uint32_t cmd_modify_lo; uint8_t cmd_reserved0[8]; struct mcx_rq_ctx cmd_rq_ctx; } __packed __aligned(4); struct mcx_cmd_modify_rq_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_destroy_rq_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint32_t cmd_rqn; uint8_t cmd_reserved1[4]; } __packed __aligned(4); struct mcx_cmd_destroy_rq_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_create_flow_table_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_flow_table_ctx { uint8_t ft_miss_action; uint8_t ft_level; uint8_t ft_reserved0; uint8_t ft_log_size; uint32_t ft_table_miss_id; uint8_t ft_reserved1[28]; } __packed __aligned(4); struct mcx_cmd_create_flow_table_mb_in { uint8_t cmd_table_type; uint8_t cmd_reserved0[7]; struct mcx_flow_table_ctx cmd_ctx; } __packed __aligned(4); struct mcx_cmd_create_flow_table_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint32_t cmd_table_id; uint8_t cmd_reserved1[4]; } __packed __aligned(4); struct mcx_cmd_destroy_flow_table_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_destroy_flow_table_mb_in { uint8_t cmd_table_type; uint8_t cmd_reserved0[3]; uint32_t cmd_table_id; uint8_t cmd_reserved1[40]; } __packed __aligned(4); struct mcx_cmd_destroy_flow_table_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_set_flow_table_root_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_set_flow_table_root_mb_in { uint8_t cmd_table_type; uint8_t cmd_reserved0[3]; uint32_t cmd_table_id; uint8_t cmd_reserved1[56]; } __packed __aligned(4); struct mcx_cmd_set_flow_table_root_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_flow_match { /* outer headers */ uint8_t mc_src_mac[6]; uint16_t mc_ethertype; uint8_t mc_dest_mac[6]; uint16_t mc_first_vlan; uint8_t mc_ip_proto; uint8_t mc_ip_dscp_ecn; uint8_t mc_vlan_flags; #define MCX_FLOW_MATCH_IP_FRAG (1 << 5) uint8_t mc_tcp_flags; uint16_t mc_tcp_sport; uint16_t mc_tcp_dport; uint32_t mc_reserved0; uint16_t mc_udp_sport; uint16_t mc_udp_dport; uint8_t mc_src_ip[16]; uint8_t mc_dest_ip[16]; /* misc parameters */ uint8_t mc_reserved1[8]; uint16_t mc_second_vlan; uint8_t mc_reserved2[2]; uint8_t mc_second_vlan_flags; uint8_t mc_reserved3[15]; uint32_t mc_outer_ipv6_flow_label; uint8_t mc_reserved4[32]; uint8_t mc_reserved[384]; } __packed __aligned(4); CTASSERT(sizeof(struct mcx_flow_match) == 512); struct mcx_cmd_create_flow_group_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_create_flow_group_mb_in { uint8_t cmd_table_type; uint8_t cmd_reserved0[3]; uint32_t cmd_table_id; uint8_t cmd_reserved1[4]; uint32_t cmd_start_flow_index; uint8_t cmd_reserved2[4]; uint32_t cmd_end_flow_index; uint8_t cmd_reserved3[23]; uint8_t cmd_match_criteria_enable; #define MCX_CREATE_FLOW_GROUP_CRIT_OUTER (1 << 0) #define MCX_CREATE_FLOW_GROUP_CRIT_MISC (1 << 1) #define MCX_CREATE_FLOW_GROUP_CRIT_INNER (1 << 2) struct mcx_flow_match cmd_match_criteria; uint8_t cmd_reserved4[448]; } __packed __aligned(4); struct mcx_cmd_create_flow_group_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint32_t cmd_group_id; uint8_t cmd_reserved1[4]; } __packed __aligned(4); struct mcx_flow_ctx { uint8_t fc_reserved0[4]; uint32_t fc_group_id; uint32_t fc_flow_tag; uint32_t fc_action; #define MCX_FLOW_CONTEXT_ACTION_ALLOW (1 << 0) #define MCX_FLOW_CONTEXT_ACTION_DROP (1 << 1) #define MCX_FLOW_CONTEXT_ACTION_FORWARD (1 << 2) #define MCX_FLOW_CONTEXT_ACTION_COUNT (1 << 3) uint32_t fc_dest_list_size; uint32_t fc_counter_list_size; uint8_t fc_reserved1[40]; struct mcx_flow_match fc_match_value; uint8_t fc_reserved2[192]; } __packed __aligned(4); #define MCX_FLOW_CONTEXT_DEST_TYPE_TABLE (1 << 24) #define MCX_FLOW_CONTEXT_DEST_TYPE_TIR (2 << 24) struct mcx_cmd_destroy_flow_group_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_destroy_flow_group_mb_in { uint8_t cmd_table_type; uint8_t cmd_reserved0[3]; uint32_t cmd_table_id; uint32_t cmd_group_id; uint8_t cmd_reserved1[36]; } __packed __aligned(4); struct mcx_cmd_destroy_flow_group_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_set_flow_table_entry_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_set_flow_table_entry_mb_in { uint8_t cmd_table_type; uint8_t cmd_reserved0[3]; uint32_t cmd_table_id; uint32_t cmd_modify_enable_mask; uint8_t cmd_reserved1[4]; uint32_t cmd_flow_index; uint8_t cmd_reserved2[28]; struct mcx_flow_ctx cmd_flow_ctx; } __packed __aligned(4); struct mcx_cmd_set_flow_table_entry_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_query_flow_table_entry_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_query_flow_table_entry_mb_in { uint8_t cmd_table_type; uint8_t cmd_reserved0[3]; uint32_t cmd_table_id; uint8_t cmd_reserved1[8]; uint32_t cmd_flow_index; uint8_t cmd_reserved2[28]; } __packed __aligned(4); struct mcx_cmd_query_flow_table_entry_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_query_flow_table_entry_mb_out { uint8_t cmd_reserved0[48]; struct mcx_flow_ctx cmd_flow_ctx; } __packed __aligned(4); struct mcx_cmd_delete_flow_table_entry_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_delete_flow_table_entry_mb_in { uint8_t cmd_table_type; uint8_t cmd_reserved0[3]; uint32_t cmd_table_id; uint8_t cmd_reserved1[8]; uint32_t cmd_flow_index; uint8_t cmd_reserved2[28]; } __packed __aligned(4); struct mcx_cmd_delete_flow_table_entry_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_query_flow_group_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_query_flow_group_mb_in { uint8_t cmd_table_type; uint8_t cmd_reserved0[3]; uint32_t cmd_table_id; uint32_t cmd_group_id; uint8_t cmd_reserved1[36]; } __packed __aligned(4); struct mcx_cmd_query_flow_group_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_query_flow_group_mb_out { uint8_t cmd_reserved0[12]; uint32_t cmd_start_flow_index; uint8_t cmd_reserved1[4]; uint32_t cmd_end_flow_index; uint8_t cmd_reserved2[20]; uint32_t cmd_match_criteria_enable; uint8_t cmd_match_criteria[512]; uint8_t cmd_reserved4[448]; } __packed __aligned(4); struct mcx_cmd_query_flow_table_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_query_flow_table_mb_in { uint8_t cmd_table_type; uint8_t cmd_reserved0[3]; uint32_t cmd_table_id; uint8_t cmd_reserved1[40]; } __packed __aligned(4); struct mcx_cmd_query_flow_table_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_query_flow_table_mb_out { uint8_t cmd_reserved0[4]; struct mcx_flow_table_ctx cmd_ctx; } __packed __aligned(4); struct mcx_cmd_alloc_flow_counter_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_query_rq_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint32_t cmd_rqn; uint8_t cmd_reserved1[4]; } __packed __aligned(4); struct mcx_cmd_query_rq_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_query_rq_mb_out { uint8_t cmd_reserved0[16]; struct mcx_rq_ctx cmd_ctx; }; struct mcx_cmd_query_sq_in { uint16_t cmd_opcode; uint8_t cmd_reserved0[4]; uint16_t cmd_op_mod; uint32_t cmd_sqn; uint8_t cmd_reserved1[4]; } __packed __aligned(4); struct mcx_cmd_query_sq_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[8]; } __packed __aligned(4); struct mcx_cmd_query_sq_mb_out { uint8_t cmd_reserved0[16]; struct mcx_sq_ctx cmd_ctx; }; struct mcx_cmd_alloc_flow_counter_out { uint8_t cmd_status; uint8_t cmd_reserved0[3]; uint32_t cmd_syndrome; uint8_t cmd_reserved1[2]; uint16_t cmd_flow_counter_id; uint8_t cmd_reserved2[4]; } __packed __aligned(4); struct mcx_wq_doorbell { uint32_t db_recv_counter; uint32_t db_send_counter; } __packed __aligned(8); struct mcx_dmamem { bus_dmamap_t mxm_map; bus_dma_segment_t mxm_seg; int mxm_nsegs; size_t mxm_size; void *mxm_kva; }; #define MCX_DMA_MAP(_mxm) ((_mxm)->mxm_map) #define MCX_DMA_DVA(_mxm) ((_mxm)->mxm_map->dm_segs[0].ds_addr) #define MCX_DMA_KVA(_mxm) ((void *)(_mxm)->mxm_kva) #define MCX_DMA_OFF(_mxm, _off) ((void *)((char *)(_mxm)->mxm_kva + (_off))) #define MCX_DMA_LEN(_mxm) ((_mxm)->mxm_size) struct mcx_hwmem { bus_dmamap_t mhm_map; bus_dma_segment_t *mhm_segs; unsigned int mhm_seg_count; unsigned int mhm_npages; }; struct mcx_slot { bus_dmamap_t ms_map; struct mbuf *ms_m; }; struct mcx_eq { int eq_n; uint32_t eq_cons; struct mcx_dmamem eq_mem; }; struct mcx_cq { int cq_n; struct mcx_dmamem cq_mem; bus_addr_t cq_doorbell; uint32_t cq_cons; uint32_t cq_count; }; struct mcx_calibration { uint64_t c_timestamp; /* previous mcx chip time */ uint64_t c_uptime; /* previous kernel nanouptime */ uint64_t c_tbase; /* mcx chip time */ uint64_t c_ubase; /* kernel nanouptime */ uint64_t c_ratio; }; #define MCX_CALIBRATE_FIRST 2 #define MCX_CALIBRATE_NORMAL 32 struct mcx_rxring { u_int rxr_total; u_int rxr_inuse; }; MBUFQ_HEAD(mcx_mbufq); struct mcx_rx { struct mcx_softc *rx_softc; int rx_rqn; struct mcx_dmamem rx_rq_mem; struct mcx_slot *rx_slots; bus_addr_t rx_doorbell; uint32_t rx_prod; callout_t rx_refill; struct mcx_rxring rx_rxr; } __aligned(64); struct mcx_tx { struct mcx_softc *tx_softc; kmutex_t tx_lock; pcq_t *tx_pcq; void *tx_softint; int tx_uar; int tx_sqn; struct mcx_dmamem tx_sq_mem; struct mcx_slot *tx_slots; bus_addr_t tx_doorbell; int tx_bf_offset; uint32_t tx_cons; uint32_t tx_prod; } __aligned(64); struct mcx_queues { void *q_ihc; struct mcx_softc *q_sc; int q_uar; int q_index; struct mcx_rx q_rx; struct mcx_tx q_tx; struct mcx_cq q_cq; struct mcx_eq q_eq; #if NKSTAT > 0 struct kstat *q_kstat; #endif }; struct mcx_flow_group { int g_id; int g_table; int g_start; int g_size; }; #define MCX_FLOW_GROUP_PROMISC 0 #define MCX_FLOW_GROUP_ALLMULTI 1 #define MCX_FLOW_GROUP_MAC 2 #define MCX_FLOW_GROUP_RSS_L4 3 #define MCX_FLOW_GROUP_RSS_L3 4 #define MCX_FLOW_GROUP_RSS_NONE 5 #define MCX_NUM_FLOW_GROUPS 6 #define MCX_HASH_SEL_L3 MCX_TIR_CTX_HASH_SEL_SRC_IP | \ MCX_TIR_CTX_HASH_SEL_DST_IP #define MCX_HASH_SEL_L4 MCX_HASH_SEL_L3 | MCX_TIR_CTX_HASH_SEL_SPORT | \ MCX_TIR_CTX_HASH_SEL_DPORT #define MCX_RSS_HASH_SEL_V4_TCP MCX_HASH_SEL_L4 | MCX_TIR_CTX_HASH_SEL_TCP |\ MCX_TIR_CTX_HASH_SEL_IPV4 #define MCX_RSS_HASH_SEL_V6_TCP MCX_HASH_SEL_L4 | MCX_TIR_CTX_HASH_SEL_TCP | \ MCX_TIR_CTX_HASH_SEL_IPV6 #define MCX_RSS_HASH_SEL_V4_UDP MCX_HASH_SEL_L4 | MCX_TIR_CTX_HASH_SEL_UDP | \ MCX_TIR_CTX_HASH_SEL_IPV4 #define MCX_RSS_HASH_SEL_V6_UDP MCX_HASH_SEL_L4 | MCX_TIR_CTX_HASH_SEL_UDP | \ MCX_TIR_CTX_HASH_SEL_IPV6 #define MCX_RSS_HASH_SEL_V4 MCX_HASH_SEL_L3 | MCX_TIR_CTX_HASH_SEL_IPV4 #define MCX_RSS_HASH_SEL_V6 MCX_HASH_SEL_L3 | MCX_TIR_CTX_HASH_SEL_IPV6 /* * There are a few different pieces involved in configuring RSS. * A Receive Queue Table (RQT) is the indirection table that maps packets to * different rx queues based on a hash value. We only create one, because * we want to scatter any traffic we can apply RSS to across all our rx * queues. Anything else will only be delivered to the first rx queue, * which doesn't require an RQT. * * A Transport Interface Receive (TIR) delivers packets to either a single rx * queue or an RQT, and in the latter case, specifies the set of fields * hashed, the hash function, and the hash key. We need one of these for each * type of RSS traffic - v4 TCP, v6 TCP, v4 UDP, v6 UDP, other v4, other v6, * and one for non-RSS traffic. * * Flow tables hold flow table entries in sequence. The first entry that * matches a packet is applied, sending the packet to either another flow * table or a TIR. We use one flow table to select packets based on * destination MAC address, and a second to apply RSS. The entries in the * first table send matching packets to the second, and the entries in the * RSS table send packets to RSS TIRs if possible, or the non-RSS TIR. * * The flow table entry that delivers packets to an RSS TIR must include match * criteria that ensure packets delivered to the TIR include all the fields * that the TIR hashes on - so for a v4 TCP TIR, the flow table entry must * only accept v4 TCP packets. Accordingly, we need flow table entries for * each TIR. * * All of this is a lot more flexible than we need, and we can describe most * of the stuff we need with a simple array. * * An RSS config creates a TIR with hashing enabled on a set of fields, * pointing to either the first rx queue or the RQT containing all the rx * queues, and a flow table entry that matches on an ether type and * optionally an ip proto, that delivers packets to the TIR. */ static struct mcx_rss_rule { int hash_sel; int flow_group; int ethertype; int ip_proto; } mcx_rss_config[] = { /* udp and tcp for v4/v6 */ { MCX_RSS_HASH_SEL_V4_TCP, MCX_FLOW_GROUP_RSS_L4, ETHERTYPE_IP, IPPROTO_TCP }, { MCX_RSS_HASH_SEL_V6_TCP, MCX_FLOW_GROUP_RSS_L4, ETHERTYPE_IPV6, IPPROTO_TCP }, { MCX_RSS_HASH_SEL_V4_UDP, MCX_FLOW_GROUP_RSS_L4, ETHERTYPE_IP, IPPROTO_UDP }, { MCX_RSS_HASH_SEL_V6_UDP, MCX_FLOW_GROUP_RSS_L4, ETHERTYPE_IPV6, IPPROTO_UDP }, /* other v4/v6 */ { MCX_RSS_HASH_SEL_V4, MCX_FLOW_GROUP_RSS_L3, ETHERTYPE_IP, 0 }, { MCX_RSS_HASH_SEL_V6, MCX_FLOW_GROUP_RSS_L3, ETHERTYPE_IPV6, 0 }, /* non v4/v6 */ { 0, MCX_FLOW_GROUP_RSS_NONE, 0, 0 } }; struct mcx_softc { device_t sc_dev; struct ethercom sc_ec; struct ifmedia sc_media; uint64_t sc_media_status; uint64_t sc_media_active; kmutex_t sc_media_mutex; pci_chipset_tag_t sc_pc; pci_intr_handle_t *sc_intrs; void *sc_ihc; pcitag_t sc_tag; bus_dma_tag_t sc_dmat; bus_space_tag_t sc_memt; bus_space_handle_t sc_memh; bus_size_t sc_mems; struct mcx_dmamem sc_cmdq_mem; unsigned int sc_cmdq_mask; unsigned int sc_cmdq_size; unsigned int sc_cmdq_token; struct mcx_hwmem sc_boot_pages; struct mcx_hwmem sc_init_pages; struct mcx_hwmem sc_regular_pages; int sc_uar; int sc_pd; int sc_tdomain; uint32_t sc_lkey; int sc_tis; int sc_tir[__arraycount(mcx_rss_config)]; int sc_rqt; struct mcx_dmamem sc_doorbell_mem; struct mcx_eq sc_admin_eq; struct mcx_eq sc_queue_eq; int sc_hardmtu; int sc_rxbufsz; int sc_bf_size; int sc_max_rqt_size; struct workqueue *sc_workq; struct work sc_port_change; int sc_mac_flow_table_id; int sc_rss_flow_table_id; struct mcx_flow_group sc_flow_group[MCX_NUM_FLOW_GROUPS]; int sc_promisc_flow_enabled; int sc_allmulti_flow_enabled; int sc_mcast_flow_base; int sc_extra_mcast; uint8_t sc_mcast_flows[MCX_NUM_MCAST_FLOWS][ETHER_ADDR_LEN]; struct mcx_calibration sc_calibration[2]; unsigned int sc_calibration_gen; callout_t sc_calibrate; uint32_t sc_mhz; uint32_t sc_khz; struct mcx_queues *sc_queues; unsigned int sc_nqueues; int sc_mcam_reg; #if NKSTAT > 0 struct kstat *sc_kstat_ieee8023; struct kstat *sc_kstat_rfc2863; struct kstat *sc_kstat_rfc2819; struct kstat *sc_kstat_rfc3635; unsigned int sc_kstat_mtmp_count; struct kstat **sc_kstat_mtmp; #endif struct timecounter sc_timecounter; }; #define DEVNAME(_sc) device_xname((_sc)->sc_dev) static int mcx_match(device_t, cfdata_t, void *); static void mcx_attach(device_t, device_t, void *); static void * mcx_establish_intr(struct mcx_softc *, int, kcpuset_t *, int (*)(void *), void *, const char *); static void mcx_rxr_init(struct mcx_rxring *, u_int, u_int); static u_int mcx_rxr_get(struct mcx_rxring *, u_int); static void mcx_rxr_put(struct mcx_rxring *, u_int); static u_int mcx_rxr_inuse(struct mcx_rxring *); #if NKSTAT > 0 static void mcx_kstat_attach(struct mcx_softc *); #endif static void mcx_timecounter_attach(struct mcx_softc *); static int mcx_version(struct mcx_softc *); static int mcx_init_wait(struct mcx_softc *); static int mcx_enable_hca(struct mcx_softc *); static int mcx_teardown_hca(struct mcx_softc *, uint16_t); static int mcx_access_hca_reg(struct mcx_softc *, uint16_t, int, void *, int); static int mcx_issi(struct mcx_softc *); static int mcx_pages(struct mcx_softc *, struct mcx_hwmem *, uint16_t); static int mcx_hca_max_caps(struct mcx_softc *); static int mcx_hca_set_caps(struct mcx_softc *); static int mcx_init_hca(struct mcx_softc *); static int mcx_set_driver_version(struct mcx_softc *); static int mcx_iff(struct mcx_softc *); static int mcx_alloc_uar(struct mcx_softc *, int *); static int mcx_alloc_pd(struct mcx_softc *); static int mcx_alloc_tdomain(struct mcx_softc *); static int mcx_create_eq(struct mcx_softc *, struct mcx_eq *, int, uint64_t, int); static int mcx_query_nic_vport_context(struct mcx_softc *, uint8_t *); static int mcx_query_special_contexts(struct mcx_softc *); static int mcx_set_port_mtu(struct mcx_softc *, int); static int mcx_create_cq(struct mcx_softc *, struct mcx_cq *, int, int, int); static int mcx_destroy_cq(struct mcx_softc *, struct mcx_cq *); static int mcx_create_sq(struct mcx_softc *, struct mcx_tx *, int, int, int); static int mcx_destroy_sq(struct mcx_softc *, struct mcx_tx *); static int mcx_ready_sq(struct mcx_softc *, struct mcx_tx *); static int mcx_create_rq(struct mcx_softc *, struct mcx_rx *, int, int); static int mcx_destroy_rq(struct mcx_softc *, struct mcx_rx *); static int mcx_ready_rq(struct mcx_softc *, struct mcx_rx *); static int mcx_create_tir_direct(struct mcx_softc *, struct mcx_rx *, int *); static int mcx_create_tir_indirect(struct mcx_softc *, int, uint32_t, int *); static int mcx_destroy_tir(struct mcx_softc *, int); static int mcx_create_tis(struct mcx_softc *, int *); static int mcx_destroy_tis(struct mcx_softc *, int); static int mcx_create_rqt(struct mcx_softc *, int, int *, int *); static int mcx_destroy_rqt(struct mcx_softc *, int); static int mcx_create_flow_table(struct mcx_softc *, int, int, int *); static int mcx_set_flow_table_root(struct mcx_softc *, int); static int mcx_destroy_flow_table(struct mcx_softc *, int); static int mcx_create_flow_group(struct mcx_softc *, int, int, int, int, int, struct mcx_flow_match *); static int mcx_destroy_flow_group(struct mcx_softc *, int); static int mcx_set_flow_table_entry_mac(struct mcx_softc *, int, int, const uint8_t *, uint32_t); static int mcx_set_flow_table_entry_proto(struct mcx_softc *, int, int, int, int, uint32_t); static int mcx_delete_flow_table_entry(struct mcx_softc *, int, int); #if NKSTAT > 0 static int mcx_query_rq(struct mcx_softc *, struct mcx_rx *, struct mcx_rq_ctx *); static int mcx_query_sq(struct mcx_softc *, struct mcx_tx *, struct mcx_sq_ctx *); static int mcx_query_cq(struct mcx_softc *, struct mcx_cq *, struct mcx_cq_ctx *); static int mcx_query_eq(struct mcx_softc *, struct mcx_eq *, struct mcx_eq_ctx *); #endif #if 0 static int mcx_dump_flow_table(struct mcx_softc *, int); static int mcx_dump_flow_table_entry(struct mcx_softc *, int, int); static int mcx_dump_flow_group(struct mcx_softc *, int); #endif /* static void mcx_cmdq_dump(const struct mcx_cmdq_entry *); static void mcx_cmdq_mbox_dump(struct mcx_dmamem *, int); */ static void mcx_refill(void *); static int mcx_process_rx(struct mcx_softc *, struct mcx_rx *, struct mcx_cq_entry *, struct mcx_mbufq *, const struct mcx_calibration *); static int mcx_process_txeof(struct mcx_softc *, struct mcx_tx *, struct mcx_cq_entry *); static void mcx_process_cq(struct mcx_softc *, struct mcx_queues *, struct mcx_cq *); static void mcx_arm_cq(struct mcx_softc *, struct mcx_cq *, int); static void mcx_arm_eq(struct mcx_softc *, struct mcx_eq *, int); static int mcx_admin_intr(void *); static int mcx_cq_intr(void *); static int mcx_init(struct ifnet *); static void mcx_stop(struct ifnet *, int); static int mcx_ioctl(struct ifnet *, u_long, void *); static void mcx_start(struct ifnet *); static int mcx_transmit(struct ifnet *, struct mbuf *); static void mcx_deferred_transmit(void *); static void mcx_media_add_types(struct mcx_softc *); static void mcx_media_status(struct ifnet *, struct ifmediareq *); static int mcx_media_change(struct ifnet *); #if 0 static int mcx_get_sffpage(struct ifnet *, struct if_sffpage *); #endif static void mcx_port_change(struct work *, void *); static void mcx_calibrate_first(struct mcx_softc *); static void mcx_calibrate(void *); static inline uint32_t mcx_rd(struct mcx_softc *, bus_size_t); static inline void mcx_wr(struct mcx_softc *, bus_size_t, uint32_t); static inline void mcx_bar(struct mcx_softc *, bus_size_t, bus_size_t, int); static uint64_t mcx_timer(struct mcx_softc *); static int mcx_dmamem_alloc(struct mcx_softc *, struct mcx_dmamem *, bus_size_t, u_int align); static void mcx_dmamem_zero(struct mcx_dmamem *); static void mcx_dmamem_free(struct mcx_softc *, struct mcx_dmamem *); static int mcx_hwmem_alloc(struct mcx_softc *, struct mcx_hwmem *, unsigned int); static void mcx_hwmem_free(struct mcx_softc *, struct mcx_hwmem *); CFATTACH_DECL_NEW(mcx, sizeof(struct mcx_softc), mcx_match, mcx_attach, NULL, NULL); static const struct { pci_vendor_id_t vendor; pci_product_id_t product; } mcx_devices[] = { { PCI_VENDOR_MELLANOX, PCI_PRODUCT_MELLANOX_MT27700 }, { PCI_VENDOR_MELLANOX, PCI_PRODUCT_MELLANOX_MT27700VF }, { PCI_VENDOR_MELLANOX, PCI_PRODUCT_MELLANOX_MT27710 }, { PCI_VENDOR_MELLANOX, PCI_PRODUCT_MELLANOX_MT27710VF }, { PCI_VENDOR_MELLANOX, PCI_PRODUCT_MELLANOX_MT27800 }, { PCI_VENDOR_MELLANOX, PCI_PRODUCT_MELLANOX_MT27800VF }, { PCI_VENDOR_MELLANOX, PCI_PRODUCT_MELLANOX_MT28800 }, { PCI_VENDOR_MELLANOX, PCI_PRODUCT_MELLANOX_MT28800VF }, { PCI_VENDOR_MELLANOX, PCI_PRODUCT_MELLANOX_MT28908 }, { PCI_VENDOR_MELLANOX, PCI_PRODUCT_MELLANOX_MT2892 }, { PCI_VENDOR_MELLANOX, PCI_PRODUCT_MELLANOX_MT2894 }, }; struct mcx_eth_proto_capability { uint64_t cap_media; uint64_t cap_baudrate; }; static const struct mcx_eth_proto_capability mcx_eth_cap_map[] = { [MCX_ETHER_CAP_SGMII] = { IFM_1000_SGMII, IF_Gbps(1) }, [MCX_ETHER_CAP_1000_KX] = { IFM_1000_KX, IF_Gbps(1) }, [MCX_ETHER_CAP_10G_CX4] = { IFM_10G_CX4, IF_Gbps(10) }, [MCX_ETHER_CAP_10G_KX4] = { IFM_10G_KX4, IF_Gbps(10) }, [MCX_ETHER_CAP_10G_KR] = { IFM_10G_KR, IF_Gbps(10) }, [MCX_ETHER_CAP_20G_KR2] = { IFM_20G_KR2, IF_Gbps(20) }, [MCX_ETHER_CAP_40G_CR4] = { IFM_40G_CR4, IF_Gbps(40) }, [MCX_ETHER_CAP_40G_KR4] = { IFM_40G_KR4, IF_Gbps(40) }, [MCX_ETHER_CAP_56G_R4] = { IFM_56G_R4, IF_Gbps(56) }, [MCX_ETHER_CAP_10G_CR] = { IFM_10G_CR1, IF_Gbps(10) }, [MCX_ETHER_CAP_10G_SR] = { IFM_10G_SR, IF_Gbps(10) }, [MCX_ETHER_CAP_10G_LR] = { IFM_10G_LR, IF_Gbps(10) }, [MCX_ETHER_CAP_40G_SR4] = { IFM_40G_SR4, IF_Gbps(40) }, [MCX_ETHER_CAP_40G_LR4] = { IFM_40G_LR4, IF_Gbps(40) }, [MCX_ETHER_CAP_50G_SR2] = { IFM_50G_SR2, IF_Gbps(50) }, [MCX_ETHER_CAP_100G_CR4] = { IFM_100G_CR4, IF_Gbps(100) }, [MCX_ETHER_CAP_100G_SR4] = { IFM_100G_SR4, IF_Gbps(100) }, [MCX_ETHER_CAP_100G_KR4] = { IFM_100G_KR4, IF_Gbps(100) }, [MCX_ETHER_CAP_100G_LR4] = { IFM_100G_LR4, IF_Gbps(100) }, [MCX_ETHER_CAP_100_TX] = { IFM_100_TX, IF_Mbps(100) }, [MCX_ETHER_CAP_1000_T] = { IFM_1000_T, IF_Gbps(1) }, [MCX_ETHER_CAP_10G_T] = { IFM_10G_T, IF_Gbps(10) }, [MCX_ETHER_CAP_25G_CR] = { IFM_25G_CR, IF_Gbps(25) }, [MCX_ETHER_CAP_25G_KR] = { IFM_25G_KR, IF_Gbps(25) }, [MCX_ETHER_CAP_25G_SR] = { IFM_25G_SR, IF_Gbps(25) }, [MCX_ETHER_CAP_50G_CR2] = { IFM_50G_CR2, IF_Gbps(50) }, [MCX_ETHER_CAP_50G_KR2] = { IFM_50G_KR2, IF_Gbps(50) }, }; static int mcx_get_id(uint32_t val) { return be32toh(val) & 0x00ffffff; } static int mcx_match(device_t parent, cfdata_t cf, void *aux) { struct pci_attach_args *pa = aux; int n; for (n = 0; n < __arraycount(mcx_devices); n++) { if (PCI_VENDOR(pa->pa_id) == mcx_devices[n].vendor && PCI_PRODUCT(pa->pa_id) == mcx_devices[n].product) return 1; } return 0; } void mcx_attach(device_t parent, device_t self, void *aux) { struct mcx_softc *sc = device_private(self); struct ifnet *ifp = &sc->sc_ec.ec_if; struct pci_attach_args *pa = aux; struct ifcapreq ifcr; uint8_t enaddr[ETHER_ADDR_LEN]; int counts[PCI_INTR_TYPE_SIZE]; char intrxname[32]; pcireg_t memtype; uint32_t r; unsigned int cq_stride; unsigned int cq_size; int i, msix; kcpuset_t *affinity; sc->sc_dev = self; sc->sc_pc = pa->pa_pc; sc->sc_tag = pa->pa_tag; if (pci_dma64_available(pa)) sc->sc_dmat = pa->pa_dmat64; else sc->sc_dmat = pa->pa_dmat; /* Map the PCI memory space */ memtype = pci_mapreg_type(sc->sc_pc, sc->sc_tag, MCX_HCA_BAR); if (pci_mapreg_map(pa, MCX_HCA_BAR, memtype, #ifdef __NetBSD__ 0, #else BUS_SPACE_MAP_PREFETCHABLE, #endif &sc->sc_memt, &sc->sc_memh, NULL, &sc->sc_mems)) { aprint_error(": unable to map register memory\n"); return; } pci_aprint_devinfo(pa, "Ethernet controller"); mutex_init(&sc->sc_media_mutex, MUTEX_DEFAULT, IPL_SOFTNET); if (mcx_version(sc) != 0) { /* error printed by mcx_version */ goto unmap; } r = mcx_rd(sc, MCX_CMDQ_ADDR_LO); cq_stride = 1 << MCX_CMDQ_LOG_STRIDE(r); /* size of the entries */ cq_size = 1 << MCX_CMDQ_LOG_SIZE(r); /* number of entries */ if (cq_size > MCX_MAX_CQE) { aprint_error_dev(self, "command queue size overflow %u\n", cq_size); goto unmap; } if (cq_stride < sizeof(struct mcx_cmdq_entry)) { aprint_error_dev(self, "command queue entry size underflow %u\n", cq_stride); goto unmap; } if (cq_stride * cq_size > MCX_PAGE_SIZE) { aprint_error_dev(self, "command queue page overflow\n"); goto unmap; } if (mcx_dmamem_alloc(sc, &sc->sc_doorbell_mem, MCX_DOORBELL_AREA_SIZE, MCX_PAGE_SIZE) != 0) { aprint_error_dev(self, "unable to allocate doorbell memory\n"); goto unmap; } if (mcx_dmamem_alloc(sc, &sc->sc_cmdq_mem, MCX_PAGE_SIZE, MCX_PAGE_SIZE) != 0) { aprint_error_dev(self, "unable to allocate command queue\n"); goto dbfree; } mcx_wr(sc, MCX_CMDQ_ADDR_HI, MCX_DMA_DVA(&sc->sc_cmdq_mem) >> 32); mcx_bar(sc, MCX_CMDQ_ADDR_HI, sizeof(uint32_t), BUS_SPACE_BARRIER_WRITE); mcx_wr(sc, MCX_CMDQ_ADDR_LO, MCX_DMA_DVA(&sc->sc_cmdq_mem)); mcx_bar(sc, MCX_CMDQ_ADDR_LO, sizeof(uint32_t), BUS_SPACE_BARRIER_WRITE); if (mcx_init_wait(sc) != 0) { aprint_error_dev(self, "timeout waiting for init\n"); goto cqfree; } sc->sc_cmdq_mask = cq_size - 1; sc->sc_cmdq_size = cq_stride; if (mcx_enable_hca(sc) != 0) { /* error printed by mcx_enable_hca */ goto cqfree; } if (mcx_issi(sc) != 0) { /* error printed by mcx_issi */ goto teardown; } if (mcx_pages(sc, &sc->sc_boot_pages, htobe16(MCX_CMD_QUERY_PAGES_BOOT)) != 0) { /* error printed by mcx_pages */ goto teardown; } if (mcx_hca_max_caps(sc) != 0) { /* error printed by mcx_hca_max_caps */ goto teardown; } if (mcx_hca_set_caps(sc) != 0) { /* error printed by mcx_hca_set_caps */ goto teardown; } if (mcx_pages(sc, &sc->sc_init_pages, htobe16(MCX_CMD_QUERY_PAGES_INIT)) != 0) { /* error printed by mcx_pages */ goto teardown; } if (mcx_init_hca(sc) != 0) { /* error printed by mcx_init_hca */ goto teardown; } if (mcx_pages(sc, &sc->sc_regular_pages, htobe16(MCX_CMD_QUERY_PAGES_REGULAR)) != 0) { /* error printed by mcx_pages */ goto teardown; } /* apparently not necessary? */ if (mcx_set_driver_version(sc) != 0) { /* error printed by mcx_set_driver_version */ goto teardown; } if (mcx_iff(sc) != 0) { /* modify nic vport context */ /* error printed by mcx_iff? */ goto teardown; } if (mcx_alloc_uar(sc, &sc->sc_uar) != 0) { /* error printed by mcx_alloc_uar */ goto teardown; } if (mcx_alloc_pd(sc) != 0) { /* error printed by mcx_alloc_pd */ goto teardown; } if (mcx_alloc_tdomain(sc) != 0) { /* error printed by mcx_alloc_tdomain */ goto teardown; } /* * PRM makes no mention of msi interrupts, just legacy and msi-x. * mellanox support tells me legacy interrupts are not supported, * so we're stuck with just msi-x. */ counts[PCI_INTR_TYPE_MSIX] = -1; counts[PCI_INTR_TYPE_MSI] = 0; counts[PCI_INTR_TYPE_INTX] = 0; if (pci_intr_alloc(pa, &sc->sc_intrs, counts, PCI_INTR_TYPE_MSIX) != 0) { aprint_error_dev(self, "unable to allocate interrupt\n"); goto teardown; } if (counts[PCI_INTR_TYPE_MSIX] < 2) { aprint_error_dev(self, "not enough MSI-X vectors\n"); goto teardown; } KASSERT(pci_intr_type(sc->sc_pc, sc->sc_intrs[0]) == PCI_INTR_TYPE_MSIX); snprintf(intrxname, sizeof(intrxname), "%s adminq", DEVNAME(sc)); sc->sc_ihc = mcx_establish_intr(sc, 0, NULL, mcx_admin_intr, sc, intrxname); if (sc->sc_ihc == NULL) { aprint_error_dev(self, "couldn't establish adminq interrupt\n"); goto teardown; } if (mcx_create_eq(sc, &sc->sc_admin_eq, sc->sc_uar, (1ull << MCX_EVENT_TYPE_INTERNAL_ERROR) | (1ull << MCX_EVENT_TYPE_PORT_CHANGE) | (1ull << MCX_EVENT_TYPE_CMD_COMPLETION) | (1ull << MCX_EVENT_TYPE_PAGE_REQUEST), 0) != 0) { /* error printed by mcx_create_eq */ goto teardown; } if (mcx_query_nic_vport_context(sc, enaddr) != 0) { /* error printed by mcx_query_nic_vport_context */ goto teardown; } if (mcx_query_special_contexts(sc) != 0) { /* error printed by mcx_query_special_contexts */ goto teardown; } if (mcx_set_port_mtu(sc, MCX_HARDMTU) != 0) { /* error printed by mcx_set_port_mtu */ goto teardown; } aprint_normal_dev(sc->sc_dev, "Ethernet address %s\n", ether_sprintf(enaddr)); msix = counts[PCI_INTR_TYPE_MSIX]; msix--; /* admin ops took one */ sc->sc_nqueues = uimin(MCX_MAX_QUEUES, msix); sc->sc_nqueues = uimin(sc->sc_nqueues, ncpu); /* Round down to a power of two. */ sc->sc_nqueues = 1U << ilog2(sc->sc_nqueues); sc->sc_queues = kmem_zalloc(sc->sc_nqueues * sizeof(*sc->sc_queues), KM_SLEEP); strlcpy(ifp->if_xname, DEVNAME(sc), IFNAMSIZ); ifp->if_softc = sc; ifp->if_flags = IFF_BROADCAST | IFF_MULTICAST | IFF_SIMPLEX; #ifdef MCX_MPSAFE ifp->if_extflags = IFEF_MPSAFE; #endif ifp->if_init = mcx_init; ifp->if_stop = mcx_stop; ifp->if_ioctl = mcx_ioctl; ifp->if_start = mcx_start; if (sc->sc_nqueues > 1) { ifp->if_transmit = mcx_transmit; } ifp->if_mtu = sc->sc_hardmtu; ifp->if_capabilities = IFCAP_CSUM_IPv4_Rx | IFCAP_CSUM_IPv4_Tx | IFCAP_CSUM_UDPv4_Rx | IFCAP_CSUM_UDPv4_Tx | IFCAP_CSUM_UDPv6_Rx | IFCAP_CSUM_UDPv6_Tx | IFCAP_CSUM_TCPv4_Rx | IFCAP_CSUM_TCPv4_Tx | IFCAP_CSUM_TCPv6_Rx | IFCAP_CSUM_TCPv6_Tx; IFQ_SET_MAXLEN(&ifp->if_snd, 1024); IFQ_SET_READY(&ifp->if_snd); sc->sc_ec.ec_capabilities = ETHERCAP_JUMBO_MTU | ETHERCAP_VLAN_MTU | ETHERCAP_VLAN_HWTAGGING; sc->sc_ec.ec_capenable |= ETHERCAP_VLAN_HWTAGGING; sc->sc_ec.ec_ifmedia = &sc->sc_media; ifmedia_init_with_lock(&sc->sc_media, IFM_IMASK, mcx_media_change, mcx_media_status, &sc->sc_media_mutex); mcx_media_add_types(sc); ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_AUTO, 0, NULL); ifmedia_set(&sc->sc_media, IFM_ETHER | IFM_AUTO); if_attach(ifp); /* Enable hardware offload by default */ memset(&ifcr, 0, sizeof(ifcr)); ifcr.ifcr_capenable = ifp->if_capabilities; ifioctl_common(ifp, SIOCSIFCAP, &ifcr); if_deferred_start_init(ifp, NULL); ether_ifattach(ifp, enaddr); kcpuset_create(&affinity, false); kcpuset_set(affinity, 0); for (i = 0; i < sc->sc_nqueues; i++) { struct mcx_queues *q = &sc->sc_queues[i]; struct mcx_rx *rx = &q->q_rx; struct mcx_tx *tx = &q->q_tx; int vec; vec = i + 1; q->q_sc = sc; q->q_index = i; if (mcx_alloc_uar(sc, &q->q_uar) != 0) { aprint_error_dev(self, "unable to alloc uar %d\n", i); goto teardown; } if (mcx_create_eq(sc, &q->q_eq, q->q_uar, 0, vec) != 0) { aprint_error_dev(self, "unable to create event queue %d\n", i); goto teardown; } rx->rx_softc = sc; callout_init(&rx->rx_refill, CALLOUT_FLAGS); callout_setfunc(&rx->rx_refill, mcx_refill, rx); tx->tx_softc = sc; mutex_init(&tx->tx_lock, MUTEX_DEFAULT, IPL_NET); tx->tx_pcq = pcq_create(MCX_TXQ_NUM, KM_SLEEP); tx->tx_softint = softint_establish(SOFTINT_NET|SOFTINT_MPSAFE, mcx_deferred_transmit, tx); snprintf(intrxname, sizeof(intrxname), "%s queue %d", DEVNAME(sc), i); q->q_ihc = mcx_establish_intr(sc, vec, affinity, mcx_cq_intr, q, intrxname); } callout_init(&sc->sc_calibrate, CALLOUT_FLAGS); callout_setfunc(&sc->sc_calibrate, mcx_calibrate, sc); if (workqueue_create(&sc->sc_workq, "mcxportchg", mcx_port_change, sc, PRI_NONE, IPL_NET, 0) != 0) { aprint_error_dev(self, "couldn't create port change workq\n"); goto teardown; } mcx_port_change(&sc->sc_port_change, sc); sc->sc_mac_flow_table_id = -1; sc->sc_rss_flow_table_id = -1; sc->sc_rqt = -1; for (i = 0; i < MCX_NUM_FLOW_GROUPS; i++) { struct mcx_flow_group *mfg = &sc->sc_flow_group[i]; mfg->g_id = -1; mfg->g_table = -1; mfg->g_size = 0; mfg->g_start = 0; } sc->sc_extra_mcast = 0; memset(sc->sc_mcast_flows, 0, sizeof(sc->sc_mcast_flows)); #if NKSTAT > 0 mcx_kstat_attach(sc); #endif mcx_timecounter_attach(sc); return; teardown: mcx_teardown_hca(sc, htobe16(MCX_CMD_TEARDOWN_HCA_GRACEFUL)); /* error printed by mcx_teardown_hca, and we're already unwinding */ cqfree: mcx_wr(sc, MCX_CMDQ_ADDR_HI, MCX_DMA_DVA(&sc->sc_cmdq_mem) >> 32); mcx_bar(sc, MCX_CMDQ_ADDR_HI, sizeof(uint64_t), BUS_SPACE_BARRIER_WRITE); mcx_wr(sc, MCX_CMDQ_ADDR_LO, MCX_DMA_DVA(&sc->sc_cmdq_mem) | MCX_CMDQ_INTERFACE_DISABLED); mcx_bar(sc, MCX_CMDQ_ADDR_LO, sizeof(uint64_t), BUS_SPACE_BARRIER_WRITE); mcx_wr(sc, MCX_CMDQ_ADDR_HI, 0); mcx_bar(sc, MCX_CMDQ_ADDR_HI, sizeof(uint64_t), BUS_SPACE_BARRIER_WRITE); mcx_wr(sc, MCX_CMDQ_ADDR_LO, MCX_CMDQ_INTERFACE_DISABLED); mcx_dmamem_free(sc, &sc->sc_cmdq_mem); dbfree: mcx_dmamem_free(sc, &sc->sc_doorbell_mem); unmap: bus_space_unmap(sc->sc_memt, sc->sc_memh, sc->sc_mems); sc->sc_mems = 0; } static void * mcx_establish_intr(struct mcx_softc *sc, int index, kcpuset_t *affinity, int (*func)(void *), void *arg, const char *xname) { char intrbuf[PCI_INTRSTR_LEN]; const char *intrstr; void *ih; pci_intr_setattr(sc->sc_pc, &sc->sc_intrs[index], PCI_INTR_MPSAFE, true); intrstr = pci_intr_string(sc->sc_pc, sc->sc_intrs[index], intrbuf, sizeof(intrbuf)); ih = pci_intr_establish_xname(sc->sc_pc, sc->sc_intrs[index], IPL_NET, func, arg, xname); if (ih == NULL) { aprint_error_dev(sc->sc_dev, "unable to establish interrupt%s%s\n", intrstr ? " at " : "", intrstr ? intrstr : ""); return NULL; } if (affinity != NULL && index > 0) { /* Round-robin affinity */ kcpuset_zero(affinity); kcpuset_set(affinity, (index - 1) % ncpu); interrupt_distribute(ih, affinity, NULL); } return ih; } static void mcx_rxr_init(struct mcx_rxring *rxr, u_int lwm __unused, u_int hwm) { rxr->rxr_total = hwm; rxr->rxr_inuse = 0; } static u_int mcx_rxr_get(struct mcx_rxring *rxr, u_int max) { const u_int taken = MIN(max, rxr->rxr_total - rxr->rxr_inuse); rxr->rxr_inuse += taken; return taken; } static void mcx_rxr_put(struct mcx_rxring *rxr, u_int n) { rxr->rxr_inuse -= n; } static u_int mcx_rxr_inuse(struct mcx_rxring *rxr) { return rxr->rxr_inuse; } static int mcx_version(struct mcx_softc *sc) { uint32_t fw0, fw1; uint16_t cmdif; fw0 = mcx_rd(sc, MCX_FW_VER); fw1 = mcx_rd(sc, MCX_CMDIF_FW_SUBVER); aprint_normal_dev(sc->sc_dev, "FW %u.%u.%04u\n", MCX_FW_VER_MAJOR(fw0), MCX_FW_VER_MINOR(fw0), MCX_FW_VER_SUBMINOR(fw1)); cmdif = MCX_CMDIF(fw1); if (cmdif != MCX_CMD_IF_SUPPORTED) { aprint_error_dev(sc->sc_dev, "unsupported command interface %u\n", cmdif); return (-1); } return (0); } static int mcx_init_wait(struct mcx_softc *sc) { unsigned int i; uint32_t r; for (i = 0; i < 2000; i++) { r = mcx_rd(sc, MCX_STATE); if ((r & MCX_STATE_MASK) == MCX_STATE_READY) return (0); delay(1000); mcx_bar(sc, MCX_STATE, sizeof(uint32_t), BUS_SPACE_BARRIER_READ); } return (-1); } static uint8_t mcx_cmdq_poll(struct mcx_softc *sc, struct mcx_cmdq_entry *cqe, unsigned int msec) { unsigned int i; for (i = 0; i < msec; i++) { bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&sc->sc_cmdq_mem), 0, MCX_DMA_LEN(&sc->sc_cmdq_mem), BUS_DMASYNC_POSTRW); if ((cqe->cq_status & MCX_CQ_STATUS_OWN_MASK) == MCX_CQ_STATUS_OWN_SW) return (0); delay(1000); } return (ETIMEDOUT); } static uint32_t mcx_mix_u64(uint32_t xor, uint64_t u64) { xor ^= u64 >> 32; xor ^= u64; return (xor); } static uint32_t mcx_mix_u32(uint32_t xor, uint32_t u32) { xor ^= u32; return (xor); } static uint32_t mcx_mix_u8(uint32_t xor, uint8_t u8) { xor ^= u8; return (xor); } static uint8_t mcx_mix_done(uint32_t xor) { xor ^= xor >> 16; xor ^= xor >> 8; return (xor); } static uint8_t mcx_xor(const void *buf, size_t len) { const uint32_t *dwords = buf; uint32_t xor = 0xff; size_t i; len /= sizeof(*dwords); for (i = 0; i < len; i++) xor ^= dwords[i]; return (mcx_mix_done(xor)); } static uint8_t mcx_cmdq_token(struct mcx_softc *sc) { uint8_t token; do { token = ++sc->sc_cmdq_token; } while (token == 0); return (token); } static void mcx_cmdq_init(struct mcx_softc *sc, struct mcx_cmdq_entry *cqe, uint32_t ilen, uint32_t olen, uint8_t token) { memset(cqe, 0, sc->sc_cmdq_size); cqe->cq_type = MCX_CMDQ_TYPE_PCIE; be32enc(&cqe->cq_input_length, ilen); be32enc(&cqe->cq_output_length, olen); cqe->cq_token = token; cqe->cq_status = MCX_CQ_STATUS_OWN_HW; } static void mcx_cmdq_sign(struct mcx_cmdq_entry *cqe) { cqe->cq_signature = ~mcx_xor(cqe, sizeof(*cqe)); } static int mcx_cmdq_verify(const struct mcx_cmdq_entry *cqe) { /* return (mcx_xor(cqe, sizeof(*cqe)) ? -1 : 0); */ return (0); } static void * mcx_cmdq_in(struct mcx_cmdq_entry *cqe) { return (&cqe->cq_input_data); } static void * mcx_cmdq_out(struct mcx_cmdq_entry *cqe) { return (&cqe->cq_output_data); } static void mcx_cmdq_post(struct mcx_softc *sc, struct mcx_cmdq_entry *cqe, unsigned int slot) { mcx_cmdq_sign(cqe); bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&sc->sc_cmdq_mem), 0, MCX_DMA_LEN(&sc->sc_cmdq_mem), BUS_DMASYNC_PRERW); mcx_wr(sc, MCX_CMDQ_DOORBELL, 1U << slot); mcx_bar(sc, MCX_CMDQ_DOORBELL, sizeof(uint32_t), BUS_SPACE_BARRIER_WRITE); } static int mcx_enable_hca(struct mcx_softc *sc) { struct mcx_cmdq_entry *cqe; struct mcx_cmd_enable_hca_in *in; struct mcx_cmd_enable_hca_out *out; int error; uint8_t status; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); mcx_cmdq_init(sc, cqe, sizeof(*in), sizeof(*out), mcx_cmdq_token(sc)); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_ENABLE_HCA); in->cmd_op_mod = htobe16(0); in->cmd_function_id = htobe16(0); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf(", hca enable timeout\n"); return (-1); } if (mcx_cmdq_verify(cqe) != 0) { printf(", hca enable command corrupt\n"); return (-1); } status = cqe->cq_output_data[0]; if (status != MCX_CQ_STATUS_OK) { printf(", hca enable failed (%x)\n", status); return (-1); } return (0); } static int mcx_teardown_hca(struct mcx_softc *sc, uint16_t profile) { struct mcx_cmdq_entry *cqe; struct mcx_cmd_teardown_hca_in *in; struct mcx_cmd_teardown_hca_out *out; int error; uint8_t status; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); mcx_cmdq_init(sc, cqe, sizeof(*in), sizeof(*out), mcx_cmdq_token(sc)); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_TEARDOWN_HCA); in->cmd_op_mod = htobe16(0); in->cmd_profile = profile; mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf(", hca teardown timeout\n"); return (-1); } if (mcx_cmdq_verify(cqe) != 0) { printf(", hca teardown command corrupt\n"); return (-1); } status = cqe->cq_output_data[0]; if (status != MCX_CQ_STATUS_OK) { printf(", hca teardown failed (%x)\n", status); return (-1); } return (0); } static int mcx_cmdq_mboxes_alloc(struct mcx_softc *sc, struct mcx_dmamem *mxm, unsigned int nmb, uint64_t *ptr, uint8_t token) { uint8_t *kva; uint64_t dva; int i; int error; error = mcx_dmamem_alloc(sc, mxm, nmb * MCX_CMDQ_MAILBOX_SIZE, MCX_CMDQ_MAILBOX_ALIGN); if (error != 0) return (error); mcx_dmamem_zero(mxm); dva = MCX_DMA_DVA(mxm); kva = MCX_DMA_KVA(mxm); for (i = 0; i < nmb; i++) { struct mcx_cmdq_mailbox *mbox = (struct mcx_cmdq_mailbox *)kva; /* patch the cqe or mbox pointing at this one */ be64enc(ptr, dva); /* fill in this mbox */ be32enc(&mbox->mb_block_number, i); mbox->mb_token = token; /* move to the next one */ ptr = &mbox->mb_next_ptr; dva += MCX_CMDQ_MAILBOX_SIZE; kva += MCX_CMDQ_MAILBOX_SIZE; } return (0); } static uint32_t mcx_cmdq_mbox_ctrl_sig(const struct mcx_cmdq_mailbox *mb) { uint32_t xor = 0xff; /* only 3 fields get set, so mix them directly */ xor = mcx_mix_u64(xor, mb->mb_next_ptr); xor = mcx_mix_u32(xor, mb->mb_block_number); xor = mcx_mix_u8(xor, mb->mb_token); return (mcx_mix_done(xor)); } static void mcx_cmdq_mboxes_sign(struct mcx_dmamem *mxm, unsigned int nmb) { uint8_t *kva; int i; kva = MCX_DMA_KVA(mxm); for (i = 0; i < nmb; i++) { struct mcx_cmdq_mailbox *mb = (struct mcx_cmdq_mailbox *)kva; uint8_t sig = mcx_cmdq_mbox_ctrl_sig(mb); mb->mb_ctrl_signature = sig; mb->mb_signature = sig ^ mcx_xor(mb->mb_data, sizeof(mb->mb_data)); kva += MCX_CMDQ_MAILBOX_SIZE; } } static void mcx_cmdq_mboxes_sync(struct mcx_softc *sc, struct mcx_dmamem *mxm, int ops) { bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(mxm), 0, MCX_DMA_LEN(mxm), ops); } static struct mcx_cmdq_mailbox * mcx_cq_mbox(struct mcx_dmamem *mxm, unsigned int i) { uint8_t *kva; kva = MCX_DMA_KVA(mxm); kva += i * MCX_CMDQ_MAILBOX_SIZE; return ((struct mcx_cmdq_mailbox *)kva); } static inline void * mcx_cq_mbox_data(struct mcx_cmdq_mailbox *mb) { return (&mb->mb_data); } static void mcx_cmdq_mboxes_copyin(struct mcx_dmamem *mxm, unsigned int nmb, void *b, size_t len) { uint8_t *buf = b; struct mcx_cmdq_mailbox *mb; int i; mb = (struct mcx_cmdq_mailbox *)MCX_DMA_KVA(mxm); for (i = 0; i < nmb; i++) { memcpy(mb->mb_data, buf, uimin(sizeof(mb->mb_data), len)); if (sizeof(mb->mb_data) >= len) break; buf += sizeof(mb->mb_data); len -= sizeof(mb->mb_data); mb++; } } static void mcx_cmdq_mboxes_pas(struct mcx_dmamem *mxm, int offset, int npages, struct mcx_dmamem *buf) { uint64_t *pas; int mbox, mbox_pages, i; mbox = offset / MCX_CMDQ_MAILBOX_DATASIZE; offset %= MCX_CMDQ_MAILBOX_DATASIZE; pas = mcx_cq_mbox_data(mcx_cq_mbox(mxm, mbox)); pas += (offset / sizeof(*pas)); mbox_pages = (MCX_CMDQ_MAILBOX_DATASIZE - offset) / sizeof(*pas); for (i = 0; i < npages; i++) { if (i == mbox_pages) { mbox++; pas = mcx_cq_mbox_data(mcx_cq_mbox(mxm, mbox)); mbox_pages += MCX_CMDQ_MAILBOX_DATASIZE / sizeof(*pas); } *pas = htobe64(MCX_DMA_DVA(buf) + (i * MCX_PAGE_SIZE)); pas++; } } static void mcx_cmdq_mboxes_copyout(struct mcx_dmamem *mxm, int nmb, void *b, size_t len) { uint8_t *buf = b; struct mcx_cmdq_mailbox *mb; int i; mb = (struct mcx_cmdq_mailbox *)MCX_DMA_KVA(mxm); for (i = 0; i < nmb; i++) { memcpy(buf, mb->mb_data, uimin(sizeof(mb->mb_data), len)); if (sizeof(mb->mb_data) >= len) break; buf += sizeof(mb->mb_data); len -= sizeof(mb->mb_data); mb++; } } static void mcx_cq_mboxes_free(struct mcx_softc *sc, struct mcx_dmamem *mxm) { mcx_dmamem_free(sc, mxm); } #if 0 static void mcx_cmdq_dump(const struct mcx_cmdq_entry *cqe) { unsigned int i; printf(" type %02x, ilen %u, iptr %016llx", cqe->cq_type, be32dec(&cqe->cq_input_length), be64dec(&cqe->cq_input_ptr)); printf(", idata "); for (i = 0; i < sizeof(cqe->cq_input_data); i++) printf("%02x", cqe->cq_input_data[i]); printf(", odata "); for (i = 0; i < sizeof(cqe->cq_output_data); i++) printf("%02x", cqe->cq_output_data[i]); printf(", optr %016llx, olen %u, token %02x, sig %02x, status %02x", be64dec(&cqe->cq_output_ptr), be32dec(&cqe->cq_output_length), cqe->cq_token, cqe->cq_signature, cqe->cq_status); } static void mcx_cmdq_mbox_dump(struct mcx_dmamem *mboxes, int num) { int i, j; uint8_t *d; for (i = 0; i < num; i++) { struct mcx_cmdq_mailbox *mbox; mbox = mcx_cq_mbox(mboxes, i); d = mcx_cq_mbox_data(mbox); for (j = 0; j < MCX_CMDQ_MAILBOX_DATASIZE; j++) { if (j != 0 && (j % 16 == 0)) printf("\n"); printf("%.2x ", d[j]); } } } #endif static int mcx_access_hca_reg(struct mcx_softc *sc, uint16_t reg, int op, void *data, int len) { struct mcx_dmamem mxm; struct mcx_cmdq_entry *cqe; struct mcx_cmd_access_reg_in *in; struct mcx_cmd_access_reg_out *out; uint8_t token = mcx_cmdq_token(sc); int error, nmb; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); mcx_cmdq_init(sc, cqe, sizeof(*in) + len, sizeof(*out) + len, token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_ACCESS_REG); in->cmd_op_mod = htobe16(op); in->cmd_register_id = htobe16(reg); nmb = howmany(len, MCX_CMDQ_MAILBOX_DATASIZE); if (mcx_cmdq_mboxes_alloc(sc, &mxm, nmb, &cqe->cq_output_ptr, token) != 0) { printf(", unable to allocate access reg mailboxen\n"); return (-1); } cqe->cq_input_ptr = cqe->cq_output_ptr; mcx_cmdq_mboxes_copyin(&mxm, nmb, data, len); mcx_cmdq_mboxes_sign(&mxm, nmb); mcx_cmdq_mboxes_sync(sc, &mxm, BUS_DMASYNC_PRERW); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); mcx_cmdq_mboxes_sync(sc, &mxm, BUS_DMASYNC_POSTRW); if (error != 0) { printf("%s: access reg (%s %x) timeout\n", DEVNAME(sc), (op == MCX_REG_OP_WRITE ? "write" : "read"), reg); goto free; } error = mcx_cmdq_verify(cqe); if (error != 0) { printf("%s: access reg (%s %x) reply corrupt\n", (op == MCX_REG_OP_WRITE ? "write" : "read"), DEVNAME(sc), reg); goto free; } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf("%s: access reg (%s %x) failed (%x, %.6x)\n", DEVNAME(sc), (op == MCX_REG_OP_WRITE ? "write" : "read"), reg, out->cmd_status, be32toh(out->cmd_syndrome)); error = -1; goto free; } mcx_cmdq_mboxes_copyout(&mxm, nmb, data, len); free: mcx_dmamem_free(sc, &mxm); return (error); } static int mcx_set_issi(struct mcx_softc *sc, struct mcx_cmdq_entry *cqe, unsigned int slot) { struct mcx_cmd_set_issi_in *in; struct mcx_cmd_set_issi_out *out; uint8_t status; mcx_cmdq_init(sc, cqe, sizeof(*in), sizeof(*out), mcx_cmdq_token(sc)); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_SET_ISSI); in->cmd_op_mod = htobe16(0); in->cmd_current_issi = htobe16(MCX_ISSI); mcx_cmdq_post(sc, cqe, slot); if (mcx_cmdq_poll(sc, cqe, 1000) != 0) return (-1); if (mcx_cmdq_verify(cqe) != 0) return (-1); status = cqe->cq_output_data[0]; if (status != MCX_CQ_STATUS_OK) return (-1); return (0); } static int mcx_issi(struct mcx_softc *sc) { struct mcx_dmamem mxm; struct mcx_cmdq_entry *cqe; struct mcx_cmd_query_issi_in *in; struct mcx_cmd_query_issi_il_out *out; struct mcx_cmd_query_issi_mb_out *mb; uint8_t token = mcx_cmdq_token(sc); uint8_t status; int error; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); mcx_cmdq_init(sc, cqe, sizeof(*in), sizeof(*out) + sizeof(*mb), token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_QUERY_ISSI); in->cmd_op_mod = htobe16(0); CTASSERT(sizeof(*mb) <= MCX_CMDQ_MAILBOX_DATASIZE); if (mcx_cmdq_mboxes_alloc(sc, &mxm, 1, &cqe->cq_output_ptr, token) != 0) { printf(", unable to allocate query issi mailbox\n"); return (-1); } mcx_cmdq_mboxes_sign(&mxm, 1); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf(", query issi timeout\n"); goto free; } error = mcx_cmdq_verify(cqe); if (error != 0) { printf(", query issi reply corrupt\n"); goto free; } status = cqe->cq_output_data[0]; switch (status) { case MCX_CQ_STATUS_OK: break; case MCX_CQ_STATUS_BAD_OPCODE: /* use ISSI 0 */ goto free; default: printf(", query issi failed (%x)\n", status); error = -1; goto free; } out = mcx_cmdq_out(cqe); if (out->cmd_current_issi == htobe16(MCX_ISSI)) { /* use ISSI 1 */ goto free; } /* don't need to read cqe anymore, can be used for SET ISSI */ mb = mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0)); CTASSERT(MCX_ISSI < NBBY); /* XXX math is hard */ if (!ISSET(mb->cmd_supported_issi[79], 1 << MCX_ISSI)) { /* use ISSI 0 */ goto free; } if (mcx_set_issi(sc, cqe, 0) != 0) { /* ignore the error, just use ISSI 0 */ } else { /* use ISSI 1 */ } free: mcx_cq_mboxes_free(sc, &mxm); return (error); } static int mcx_query_pages(struct mcx_softc *sc, uint16_t type, int32_t *npages, uint16_t *func_id) { struct mcx_cmdq_entry *cqe; struct mcx_cmd_query_pages_in *in; struct mcx_cmd_query_pages_out *out; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); mcx_cmdq_init(sc, cqe, sizeof(*in), sizeof(*out), mcx_cmdq_token(sc)); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_QUERY_PAGES); in->cmd_op_mod = type; mcx_cmdq_post(sc, cqe, 0); if (mcx_cmdq_poll(sc, cqe, 1000) != 0) { printf(", query pages timeout\n"); return (-1); } if (mcx_cmdq_verify(cqe) != 0) { printf(", query pages reply corrupt\n"); return (-1); } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf(", query pages failed (%x)\n", out->cmd_status); return (-1); } *func_id = out->cmd_func_id; *npages = be32dec(&out->cmd_num_pages); return (0); } struct bus_dma_iter { bus_dmamap_t i_map; bus_size_t i_offset; unsigned int i_index; }; static void bus_dma_iter_init(struct bus_dma_iter *i, bus_dmamap_t map) { i->i_map = map; i->i_offset = 0; i->i_index = 0; } static bus_addr_t bus_dma_iter_addr(struct bus_dma_iter *i) { return (i->i_map->dm_segs[i->i_index].ds_addr + i->i_offset); } static void bus_dma_iter_add(struct bus_dma_iter *i, bus_size_t size) { bus_dma_segment_t *seg = i->i_map->dm_segs + i->i_index; bus_size_t diff; do { diff = seg->ds_len - i->i_offset; if (size < diff) break; size -= diff; seg++; i->i_offset = 0; i->i_index++; } while (size > 0); i->i_offset += size; } static int mcx_add_pages(struct mcx_softc *sc, struct mcx_hwmem *mhm, uint16_t func_id) { struct mcx_dmamem mxm; struct mcx_cmdq_entry *cqe; struct mcx_cmd_manage_pages_in *in; struct mcx_cmd_manage_pages_out *out; unsigned int paslen, nmb, i, j, npages; struct bus_dma_iter iter; uint64_t *pas; uint8_t status; uint8_t token = mcx_cmdq_token(sc); int error; npages = mhm->mhm_npages; paslen = sizeof(*pas) * npages; nmb = howmany(paslen, MCX_CMDQ_MAILBOX_DATASIZE); cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); mcx_cmdq_init(sc, cqe, sizeof(*in) + paslen, sizeof(*out), token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_MANAGE_PAGES); in->cmd_op_mod = htobe16(MCX_CMD_MANAGE_PAGES_ALLOC_SUCCESS); in->cmd_func_id = func_id; be32enc(&in->cmd_input_num_entries, npages); if (mcx_cmdq_mboxes_alloc(sc, &mxm, nmb, &cqe->cq_input_ptr, token) != 0) { printf(", unable to allocate manage pages mailboxen\n"); return (-1); } bus_dma_iter_init(&iter, mhm->mhm_map); for (i = 0; i < nmb; i++) { unsigned int lim; pas = mcx_cq_mbox_data(mcx_cq_mbox(&mxm, i)); lim = uimin(MCX_CMDQ_MAILBOX_DATASIZE / sizeof(*pas), npages); for (j = 0; j < lim; j++) { be64enc(&pas[j], bus_dma_iter_addr(&iter)); bus_dma_iter_add(&iter, MCX_PAGE_SIZE); } npages -= lim; } mcx_cmdq_mboxes_sign(&mxm, nmb); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf(", manage pages timeout\n"); goto free; } error = mcx_cmdq_verify(cqe); if (error != 0) { printf(", manage pages reply corrupt\n"); goto free; } status = cqe->cq_output_data[0]; if (status != MCX_CQ_STATUS_OK) { printf(", manage pages failed (%x)\n", status); error = -1; goto free; } free: mcx_dmamem_free(sc, &mxm); return (error); } static int mcx_pages(struct mcx_softc *sc, struct mcx_hwmem *mhm, uint16_t type) { int32_t npages; uint16_t func_id; if (mcx_query_pages(sc, type, &npages, &func_id) != 0) { /* error printed by mcx_query_pages */ return (-1); } if (npages < 1) return (0); if (mcx_hwmem_alloc(sc, mhm, npages) != 0) { printf(", unable to allocate hwmem\n"); return (-1); } if (mcx_add_pages(sc, mhm, func_id) != 0) { printf(", unable to add hwmem\n"); goto free; } return (0); free: mcx_hwmem_free(sc, mhm); return (-1); } static int mcx_hca_max_caps(struct mcx_softc *sc) { struct mcx_dmamem mxm; struct mcx_cmdq_entry *cqe; struct mcx_cmd_query_hca_cap_in *in; struct mcx_cmd_query_hca_cap_out *out; struct mcx_cmdq_mailbox *mb; struct mcx_cap_device *hca; uint8_t status; uint8_t token = mcx_cmdq_token(sc); int error; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); mcx_cmdq_init(sc, cqe, sizeof(*in), sizeof(*out) + MCX_HCA_CAP_LEN, token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_QUERY_HCA_CAP); in->cmd_op_mod = htobe16(MCX_CMD_QUERY_HCA_CAP_MAX | MCX_CMD_QUERY_HCA_CAP_DEVICE); if (mcx_cmdq_mboxes_alloc(sc, &mxm, MCX_HCA_CAP_NMAILBOXES, &cqe->cq_output_ptr, token) != 0) { printf(", unable to allocate query hca caps mailboxen\n"); return (-1); } mcx_cmdq_mboxes_sign(&mxm, MCX_HCA_CAP_NMAILBOXES); mcx_cmdq_mboxes_sync(sc, &mxm, BUS_DMASYNC_PRERW); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); mcx_cmdq_mboxes_sync(sc, &mxm, BUS_DMASYNC_POSTRW); if (error != 0) { printf(", query hca caps timeout\n"); goto free; } error = mcx_cmdq_verify(cqe); if (error != 0) { printf(", query hca caps reply corrupt\n"); goto free; } status = cqe->cq_output_data[0]; if (status != MCX_CQ_STATUS_OK) { printf(", query hca caps failed (%x)\n", status); error = -1; goto free; } mb = mcx_cq_mbox(&mxm, 0); hca = mcx_cq_mbox_data(mb); if ((hca->port_type & MCX_CAP_DEVICE_PORT_TYPE) != MCX_CAP_DEVICE_PORT_TYPE_ETH) { printf(", not in ethernet mode\n"); error = -1; goto free; } if (hca->log_pg_sz > PAGE_SHIFT) { printf(", minimum system page shift %u is too large\n", hca->log_pg_sz); error = -1; goto free; } /* * blueflame register is split into two buffers, and we must alternate * between the two of them. */ sc->sc_bf_size = (1 << hca->log_bf_reg_size) / 2; sc->sc_max_rqt_size = (1 << hca->log_max_rqt_size); if (hca->local_ca_ack_delay & MCX_CAP_DEVICE_MCAM_REG) sc->sc_mcam_reg = 1; sc->sc_mhz = be32dec(&hca->device_frequency_mhz); sc->sc_khz = be32dec(&hca->device_frequency_khz); free: mcx_dmamem_free(sc, &mxm); return (error); } static int mcx_hca_set_caps(struct mcx_softc *sc) { struct mcx_dmamem mxm; struct mcx_cmdq_entry *cqe; struct mcx_cmd_query_hca_cap_in *in; struct mcx_cmd_query_hca_cap_out *out; struct mcx_cmdq_mailbox *mb; struct mcx_cap_device *hca; uint8_t status; uint8_t token = mcx_cmdq_token(sc); int error; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); mcx_cmdq_init(sc, cqe, sizeof(*in), sizeof(*out) + MCX_HCA_CAP_LEN, token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_QUERY_HCA_CAP); in->cmd_op_mod = htobe16(MCX_CMD_QUERY_HCA_CAP_CURRENT | MCX_CMD_QUERY_HCA_CAP_DEVICE); if (mcx_cmdq_mboxes_alloc(sc, &mxm, MCX_HCA_CAP_NMAILBOXES, &cqe->cq_output_ptr, token) != 0) { printf(", unable to allocate manage pages mailboxen\n"); return (-1); } mcx_cmdq_mboxes_sign(&mxm, MCX_HCA_CAP_NMAILBOXES); mcx_cmdq_mboxes_sync(sc, &mxm, BUS_DMASYNC_PRERW); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); mcx_cmdq_mboxes_sync(sc, &mxm, BUS_DMASYNC_POSTRW); if (error != 0) { printf(", query hca caps timeout\n"); goto free; } error = mcx_cmdq_verify(cqe); if (error != 0) { printf(", query hca caps reply corrupt\n"); goto free; } status = cqe->cq_output_data[0]; if (status != MCX_CQ_STATUS_OK) { printf(", query hca caps failed (%x)\n", status); error = -1; goto free; } mb = mcx_cq_mbox(&mxm, 0); hca = mcx_cq_mbox_data(mb); hca->log_pg_sz = PAGE_SHIFT; free: mcx_dmamem_free(sc, &mxm); return (error); } static int mcx_init_hca(struct mcx_softc *sc) { struct mcx_cmdq_entry *cqe; struct mcx_cmd_init_hca_in *in; struct mcx_cmd_init_hca_out *out; int error; uint8_t status; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); mcx_cmdq_init(sc, cqe, sizeof(*in), sizeof(*out), mcx_cmdq_token(sc)); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_INIT_HCA); in->cmd_op_mod = htobe16(0); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf(", hca init timeout\n"); return (-1); } if (mcx_cmdq_verify(cqe) != 0) { printf(", hca init command corrupt\n"); return (-1); } status = cqe->cq_output_data[0]; if (status != MCX_CQ_STATUS_OK) { printf(", hca init failed (%x)\n", status); return (-1); } return (0); } static int mcx_set_driver_version(struct mcx_softc *sc) { struct mcx_dmamem mxm; struct mcx_cmdq_entry *cqe; struct mcx_cmd_set_driver_version_in *in; struct mcx_cmd_set_driver_version_out *out; int error; int token; uint8_t status; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); token = mcx_cmdq_token(sc); mcx_cmdq_init(sc, cqe, sizeof(*in) + sizeof(struct mcx_cmd_set_driver_version), sizeof(*out), token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_SET_DRIVER_VERSION); in->cmd_op_mod = htobe16(0); if (mcx_cmdq_mboxes_alloc(sc, &mxm, 1, &cqe->cq_input_ptr, token) != 0) { printf(", unable to allocate set driver version mailboxen\n"); return (-1); } strlcpy(mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0)), "OpenBSD,mcx,1.000.000000", MCX_CMDQ_MAILBOX_DATASIZE); mcx_cmdq_mboxes_sign(&mxm, 1); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf(", set driver version timeout\n"); goto free; } if (mcx_cmdq_verify(cqe) != 0) { printf(", set driver version command corrupt\n"); goto free; } status = cqe->cq_output_data[0]; if (status != MCX_CQ_STATUS_OK) { printf(", set driver version failed (%x)\n", status); error = -1; goto free; } free: mcx_dmamem_free(sc, &mxm); return (error); } static int mcx_iff(struct mcx_softc *sc) { struct ifnet *ifp = &sc->sc_ec.ec_if; struct mcx_dmamem mxm; struct mcx_cmdq_entry *cqe; struct mcx_cmd_modify_nic_vport_context_in *in; struct mcx_cmd_modify_nic_vport_context_out *out; struct mcx_nic_vport_ctx *ctx; int error; int token; int insize; uint32_t dest; dest = MCX_FLOW_CONTEXT_DEST_TYPE_TABLE | sc->sc_rss_flow_table_id; /* enable or disable the promisc flow */ if (ISSET(ifp->if_flags, IFF_PROMISC)) { if (sc->sc_promisc_flow_enabled == 0) { mcx_set_flow_table_entry_mac(sc, MCX_FLOW_GROUP_PROMISC, 0, NULL, dest); sc->sc_promisc_flow_enabled = 1; } } else if (sc->sc_promisc_flow_enabled != 0) { mcx_delete_flow_table_entry(sc, MCX_FLOW_GROUP_PROMISC, 0); sc->sc_promisc_flow_enabled = 0; } /* enable or disable the all-multicast flow */ if (ISSET(ifp->if_flags, IFF_ALLMULTI)) { if (sc->sc_allmulti_flow_enabled == 0) { uint8_t mcast[ETHER_ADDR_LEN]; memset(mcast, 0, sizeof(mcast)); mcast[0] = 0x01; mcx_set_flow_table_entry_mac(sc, MCX_FLOW_GROUP_ALLMULTI, 0, mcast, dest); sc->sc_allmulti_flow_enabled = 1; } } else if (sc->sc_allmulti_flow_enabled != 0) { mcx_delete_flow_table_entry(sc, MCX_FLOW_GROUP_ALLMULTI, 0); sc->sc_allmulti_flow_enabled = 0; } insize = sizeof(struct mcx_nic_vport_ctx) + 240; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); token = mcx_cmdq_token(sc); mcx_cmdq_init(sc, cqe, sizeof(*in) + insize, sizeof(*out), token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_MODIFY_NIC_VPORT_CONTEXT); in->cmd_op_mod = htobe16(0); in->cmd_field_select = htobe32( MCX_CMD_MODIFY_NIC_VPORT_CONTEXT_FIELD_PROMISC | MCX_CMD_MODIFY_NIC_VPORT_CONTEXT_FIELD_MTU); if (mcx_cmdq_mboxes_alloc(sc, &mxm, 1, &cqe->cq_input_ptr, token) != 0) { printf(", unable to allocate modify " "nic vport context mailboxen\n"); return (-1); } ctx = (struct mcx_nic_vport_ctx *) (((char *)mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0))) + 240); ctx->vp_mtu = htobe32(sc->sc_hardmtu); /* * always leave promisc-all enabled on the vport since we * can't give it a vlan list, and we're already doing multicast * filtering in the flow table. */ ctx->vp_flags = htobe16(MCX_NIC_VPORT_CTX_PROMISC_ALL); mcx_cmdq_mboxes_sign(&mxm, 1); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf(", modify nic vport context timeout\n"); goto free; } if (mcx_cmdq_verify(cqe) != 0) { printf(", modify nic vport context command corrupt\n"); goto free; } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf(", modify nic vport context failed (%x, %x)\n", out->cmd_status, be32toh(out->cmd_syndrome)); error = -1; goto free; } free: mcx_dmamem_free(sc, &mxm); return (error); } static int mcx_alloc_uar(struct mcx_softc *sc, int *uar) { struct mcx_cmdq_entry *cqe; struct mcx_cmd_alloc_uar_in *in; struct mcx_cmd_alloc_uar_out *out; int error; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); mcx_cmdq_init(sc, cqe, sizeof(*in), sizeof(*out), mcx_cmdq_token(sc)); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_ALLOC_UAR); in->cmd_op_mod = htobe16(0); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf(", alloc uar timeout\n"); return (-1); } if (mcx_cmdq_verify(cqe) != 0) { printf(", alloc uar command corrupt\n"); return (-1); } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf(", alloc uar failed (%x)\n", out->cmd_status); return (-1); } *uar = mcx_get_id(out->cmd_uar); return (0); } static int mcx_create_eq(struct mcx_softc *sc, struct mcx_eq *eq, int uar, uint64_t events, int vector) { struct mcx_cmdq_entry *cqe; struct mcx_dmamem mxm; struct mcx_cmd_create_eq_in *in; struct mcx_cmd_create_eq_mb_in *mbin; struct mcx_cmd_create_eq_out *out; struct mcx_eq_entry *eqe; int error; uint64_t *pas; int insize, npages, paslen, i, token; eq->eq_cons = 0; npages = howmany((1 << MCX_LOG_EQ_SIZE) * sizeof(struct mcx_eq_entry), MCX_PAGE_SIZE); paslen = npages * sizeof(*pas); insize = sizeof(struct mcx_cmd_create_eq_mb_in) + paslen; if (mcx_dmamem_alloc(sc, &eq->eq_mem, npages * MCX_PAGE_SIZE, MCX_PAGE_SIZE) != 0) { printf(", unable to allocate event queue memory\n"); return (-1); } eqe = (struct mcx_eq_entry *)MCX_DMA_KVA(&eq->eq_mem); for (i = 0; i < (1 << MCX_LOG_EQ_SIZE); i++) { eqe[i].eq_owner = MCX_EQ_ENTRY_OWNER_INIT; } cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); token = mcx_cmdq_token(sc); mcx_cmdq_init(sc, cqe, sizeof(*in) + insize, sizeof(*out), token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_CREATE_EQ); in->cmd_op_mod = htobe16(0); if (mcx_cmdq_mboxes_alloc(sc, &mxm, howmany(insize, MCX_CMDQ_MAILBOX_DATASIZE), &cqe->cq_input_ptr, token) != 0) { printf(", unable to allocate create eq mailboxen\n"); goto free_eq; } mbin = mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0)); mbin->cmd_eq_ctx.eq_uar_size = htobe32( (MCX_LOG_EQ_SIZE << MCX_EQ_CTX_LOG_EQ_SIZE_SHIFT) | uar); mbin->cmd_eq_ctx.eq_intr = vector; mbin->cmd_event_bitmask = htobe64(events); bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&eq->eq_mem), 0, MCX_DMA_LEN(&eq->eq_mem), BUS_DMASYNC_PREREAD); /* physical addresses follow the mailbox in data */ mcx_cmdq_mboxes_pas(&mxm, sizeof(*mbin), npages, &eq->eq_mem); mcx_cmdq_mboxes_sign(&mxm, howmany(insize, MCX_CMDQ_MAILBOX_DATASIZE)); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf(", create eq timeout\n"); goto free_mxm; } if (mcx_cmdq_verify(cqe) != 0) { printf(", create eq command corrupt\n"); goto free_mxm; } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf(", create eq failed (%x, %x)\n", out->cmd_status, be32toh(out->cmd_syndrome)); goto free_mxm; } eq->eq_n = mcx_get_id(out->cmd_eqn); mcx_dmamem_free(sc, &mxm); mcx_arm_eq(sc, eq, uar); return (0); free_mxm: bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&eq->eq_mem), 0, MCX_DMA_LEN(&eq->eq_mem), BUS_DMASYNC_POSTREAD); mcx_dmamem_free(sc, &mxm); free_eq: mcx_dmamem_free(sc, &eq->eq_mem); return (-1); } static int mcx_alloc_pd(struct mcx_softc *sc) { struct mcx_cmdq_entry *cqe; struct mcx_cmd_alloc_pd_in *in; struct mcx_cmd_alloc_pd_out *out; int error; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); mcx_cmdq_init(sc, cqe, sizeof(*in), sizeof(*out), mcx_cmdq_token(sc)); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_ALLOC_PD); in->cmd_op_mod = htobe16(0); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf(", alloc pd timeout\n"); return (-1); } if (mcx_cmdq_verify(cqe) != 0) { printf(", alloc pd command corrupt\n"); return (-1); } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf(", alloc pd failed (%x)\n", out->cmd_status); return (-1); } sc->sc_pd = mcx_get_id(out->cmd_pd); return (0); } static int mcx_alloc_tdomain(struct mcx_softc *sc) { struct mcx_cmdq_entry *cqe; struct mcx_cmd_alloc_td_in *in; struct mcx_cmd_alloc_td_out *out; int error; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); mcx_cmdq_init(sc, cqe, sizeof(*in), sizeof(*out), mcx_cmdq_token(sc)); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_ALLOC_TRANSPORT_DOMAIN); in->cmd_op_mod = htobe16(0); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf(", alloc transport domain timeout\n"); return (-1); } if (mcx_cmdq_verify(cqe) != 0) { printf(", alloc transport domain command corrupt\n"); return (-1); } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf(", alloc transport domain failed (%x)\n", out->cmd_status); return (-1); } sc->sc_tdomain = mcx_get_id(out->cmd_tdomain); return (0); } static int mcx_query_nic_vport_context(struct mcx_softc *sc, uint8_t *enaddr) { struct mcx_dmamem mxm; struct mcx_cmdq_entry *cqe; struct mcx_cmd_query_nic_vport_context_in *in; struct mcx_cmd_query_nic_vport_context_out *out; struct mcx_nic_vport_ctx *ctx; uint8_t *addr; int error, token, i; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); token = mcx_cmdq_token(sc); mcx_cmdq_init(sc, cqe, sizeof(*in), sizeof(*out) + sizeof(*ctx), token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_QUERY_NIC_VPORT_CONTEXT); in->cmd_op_mod = htobe16(0); in->cmd_allowed_list_type = 0; if (mcx_cmdq_mboxes_alloc(sc, &mxm, 1, &cqe->cq_output_ptr, token) != 0) { printf(", unable to allocate " "query nic vport context mailboxen\n"); return (-1); } mcx_cmdq_mboxes_sign(&mxm, 1); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf(", query nic vport context timeout\n"); goto free; } if (mcx_cmdq_verify(cqe) != 0) { printf(", query nic vport context command corrupt\n"); goto free; } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf(", query nic vport context failed (%x, %x)\n", out->cmd_status, be32toh(out->cmd_syndrome)); error = -1; goto free; } ctx = (struct mcx_nic_vport_ctx *) mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0)); addr = (uint8_t *)&ctx->vp_perm_addr; for (i = 0; i < ETHER_ADDR_LEN; i++) { enaddr[i] = addr[i + 2]; } free: mcx_dmamem_free(sc, &mxm); return (error); } static int mcx_query_special_contexts(struct mcx_softc *sc) { struct mcx_cmdq_entry *cqe; struct mcx_cmd_query_special_ctx_in *in; struct mcx_cmd_query_special_ctx_out *out; int error; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); mcx_cmdq_init(sc, cqe, sizeof(*in), sizeof(*out), mcx_cmdq_token(sc)); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_QUERY_SPECIAL_CONTEXTS); in->cmd_op_mod = htobe16(0); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf(", query special contexts timeout\n"); return (-1); } if (mcx_cmdq_verify(cqe) != 0) { printf(", query special contexts command corrupt\n"); return (-1); } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf(", query special contexts failed (%x)\n", out->cmd_status); return (-1); } sc->sc_lkey = be32toh(out->cmd_resd_lkey); return (0); } static int mcx_set_port_mtu(struct mcx_softc *sc, int mtu) { struct mcx_reg_pmtu pmtu; int error; /* read max mtu */ memset(&pmtu, 0, sizeof(pmtu)); pmtu.rp_local_port = 1; error = mcx_access_hca_reg(sc, MCX_REG_PMTU, MCX_REG_OP_READ, &pmtu, sizeof(pmtu)); if (error != 0) { printf(", unable to get port MTU\n"); return error; } mtu = uimin(mtu, be16toh(pmtu.rp_max_mtu)); pmtu.rp_admin_mtu = htobe16(mtu); error = mcx_access_hca_reg(sc, MCX_REG_PMTU, MCX_REG_OP_WRITE, &pmtu, sizeof(pmtu)); if (error != 0) { printf(", unable to set port MTU\n"); return error; } sc->sc_hardmtu = mtu; sc->sc_rxbufsz = roundup(mtu + ETHER_ALIGN, sizeof(long)); return 0; } static int mcx_create_cq(struct mcx_softc *sc, struct mcx_cq *cq, int uar, int db, int eqn) { struct mcx_cmdq_entry *cmde; struct mcx_cq_entry *cqe; struct mcx_dmamem mxm; struct mcx_cmd_create_cq_in *in; struct mcx_cmd_create_cq_mb_in *mbin; struct mcx_cmd_create_cq_out *out; int error; uint64_t *pas; int insize, npages, paslen, i, token; cq->cq_doorbell = MCX_CQ_DOORBELL_BASE + (MCX_CQ_DOORBELL_STRIDE * db); npages = howmany((1 << MCX_LOG_CQ_SIZE) * sizeof(struct mcx_cq_entry), MCX_PAGE_SIZE); paslen = npages * sizeof(*pas); insize = sizeof(struct mcx_cmd_create_cq_mb_in) + paslen; if (mcx_dmamem_alloc(sc, &cq->cq_mem, npages * MCX_PAGE_SIZE, MCX_PAGE_SIZE) != 0) { printf("%s: unable to allocate completion queue memory\n", DEVNAME(sc)); return (-1); } cqe = MCX_DMA_KVA(&cq->cq_mem); for (i = 0; i < (1 << MCX_LOG_CQ_SIZE); i++) { cqe[i].cq_opcode_owner = MCX_CQ_ENTRY_FLAG_OWNER; } cmde = MCX_DMA_KVA(&sc->sc_cmdq_mem); token = mcx_cmdq_token(sc); mcx_cmdq_init(sc, cmde, sizeof(*in) + insize, sizeof(*out), token); in = mcx_cmdq_in(cmde); in->cmd_opcode = htobe16(MCX_CMD_CREATE_CQ); in->cmd_op_mod = htobe16(0); if (mcx_cmdq_mboxes_alloc(sc, &mxm, howmany(insize, MCX_CMDQ_MAILBOX_DATASIZE), &cmde->cq_input_ptr, token) != 0) { printf("%s: unable to allocate create cq mailboxen\n", DEVNAME(sc)); goto free_cq; } mbin = mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0)); mbin->cmd_cq_ctx.cq_uar_size = htobe32( (MCX_LOG_CQ_SIZE << MCX_CQ_CTX_LOG_CQ_SIZE_SHIFT) | uar); mbin->cmd_cq_ctx.cq_eqn = htobe32(eqn); mbin->cmd_cq_ctx.cq_period_max_count = htobe32( (MCX_CQ_MOD_PERIOD << MCX_CQ_CTX_PERIOD_SHIFT) | MCX_CQ_MOD_COUNTER); mbin->cmd_cq_ctx.cq_doorbell = htobe64( MCX_DMA_DVA(&sc->sc_doorbell_mem) + cq->cq_doorbell); bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&cq->cq_mem), 0, MCX_DMA_LEN(&cq->cq_mem), BUS_DMASYNC_PREREAD); /* physical addresses follow the mailbox in data */ mcx_cmdq_mboxes_pas(&mxm, sizeof(*mbin), npages, &cq->cq_mem); mcx_cmdq_post(sc, cmde, 0); error = mcx_cmdq_poll(sc, cmde, 1000); if (error != 0) { printf("%s: create cq timeout\n", DEVNAME(sc)); goto free_mxm; } if (mcx_cmdq_verify(cmde) != 0) { printf("%s: create cq command corrupt\n", DEVNAME(sc)); goto free_mxm; } out = mcx_cmdq_out(cmde); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf("%s: create cq failed (%x, %x)\n", DEVNAME(sc), out->cmd_status, be32toh(out->cmd_syndrome)); goto free_mxm; } cq->cq_n = mcx_get_id(out->cmd_cqn); cq->cq_cons = 0; cq->cq_count = 0; mcx_dmamem_free(sc, &mxm); bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&sc->sc_doorbell_mem), cq->cq_doorbell, sizeof(struct mcx_cq_doorbell), BUS_DMASYNC_PREWRITE); mcx_arm_cq(sc, cq, uar); return (0); free_mxm: bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&cq->cq_mem), 0, MCX_DMA_LEN(&cq->cq_mem), BUS_DMASYNC_POSTREAD); mcx_dmamem_free(sc, &mxm); free_cq: mcx_dmamem_free(sc, &cq->cq_mem); return (-1); } static int mcx_destroy_cq(struct mcx_softc *sc, struct mcx_cq *cq) { struct mcx_cmdq_entry *cqe; struct mcx_cmd_destroy_cq_in *in; struct mcx_cmd_destroy_cq_out *out; int error; int token; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); token = mcx_cmdq_token(sc); mcx_cmdq_init(sc, cqe, sizeof(*in), sizeof(*out), token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_DESTROY_CQ); in->cmd_op_mod = htobe16(0); in->cmd_cqn = htobe32(cq->cq_n); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: destroy cq timeout\n", DEVNAME(sc)); return error; } if (mcx_cmdq_verify(cqe) != 0) { printf("%s: destroy cq command corrupt\n", DEVNAME(sc)); return error; } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf("%s: destroy cq failed (%x, %x)\n", DEVNAME(sc), out->cmd_status, be32toh(out->cmd_syndrome)); return -1; } bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&sc->sc_doorbell_mem), cq->cq_doorbell, sizeof(struct mcx_cq_doorbell), BUS_DMASYNC_POSTWRITE); bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&cq->cq_mem), 0, MCX_DMA_LEN(&cq->cq_mem), BUS_DMASYNC_POSTREAD); mcx_dmamem_free(sc, &cq->cq_mem); cq->cq_n = 0; cq->cq_cons = 0; cq->cq_count = 0; return 0; } static int mcx_create_rq(struct mcx_softc *sc, struct mcx_rx *rx, int db, int cqn) { struct mcx_cmdq_entry *cqe; struct mcx_dmamem mxm; struct mcx_cmd_create_rq_in *in; struct mcx_cmd_create_rq_out *out; struct mcx_rq_ctx *mbin; int error; uint64_t *pas; uint32_t rq_flags; int insize, npages, paslen, token; rx->rx_doorbell = MCX_WQ_DOORBELL_BASE + (db * MCX_WQ_DOORBELL_STRIDE); npages = howmany((1 << MCX_LOG_RQ_SIZE) * sizeof(struct mcx_rq_entry), MCX_PAGE_SIZE); paslen = npages * sizeof(*pas); insize = 0x10 + sizeof(struct mcx_rq_ctx) + paslen; if (mcx_dmamem_alloc(sc, &rx->rx_rq_mem, npages * MCX_PAGE_SIZE, MCX_PAGE_SIZE) != 0) { printf("%s: unable to allocate receive queue memory\n", DEVNAME(sc)); return (-1); } cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); token = mcx_cmdq_token(sc); mcx_cmdq_init(sc, cqe, sizeof(*in) + insize, sizeof(*out), token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_CREATE_RQ); in->cmd_op_mod = htobe16(0); if (mcx_cmdq_mboxes_alloc(sc, &mxm, howmany(insize, MCX_CMDQ_MAILBOX_DATASIZE), &cqe->cq_input_ptr, token) != 0) { printf("%s: unable to allocate create rq mailboxen\n", DEVNAME(sc)); goto free_rq; } mbin = (struct mcx_rq_ctx *) (((char *)mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0))) + 0x10); rq_flags = MCX_RQ_CTX_RLKEY; mbin->rq_flags = htobe32(rq_flags); mbin->rq_cqn = htobe32(cqn); mbin->rq_wq.wq_type = MCX_WQ_CTX_TYPE_CYCLIC; mbin->rq_wq.wq_pd = htobe32(sc->sc_pd); mbin->rq_wq.wq_doorbell = htobe64(MCX_DMA_DVA(&sc->sc_doorbell_mem) + rx->rx_doorbell); mbin->rq_wq.wq_log_stride = htobe16(4); mbin->rq_wq.wq_log_size = MCX_LOG_RQ_SIZE; bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&rx->rx_rq_mem), 0, MCX_DMA_LEN(&rx->rx_rq_mem), BUS_DMASYNC_PREWRITE); /* physical addresses follow the mailbox in data */ mcx_cmdq_mboxes_pas(&mxm, sizeof(*mbin) + 0x10, npages, &rx->rx_rq_mem); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: create rq timeout\n", DEVNAME(sc)); goto free_mxm; } if (mcx_cmdq_verify(cqe) != 0) { printf("%s: create rq command corrupt\n", DEVNAME(sc)); goto free_mxm; } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf("%s: create rq failed (%x, %x)\n", DEVNAME(sc), out->cmd_status, be32toh(out->cmd_syndrome)); goto free_mxm; } rx->rx_rqn = mcx_get_id(out->cmd_rqn); mcx_dmamem_free(sc, &mxm); bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&sc->sc_doorbell_mem), rx->rx_doorbell, sizeof(uint32_t), BUS_DMASYNC_PREWRITE); return (0); free_mxm: bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&rx->rx_rq_mem), 0, MCX_DMA_LEN(&rx->rx_rq_mem), BUS_DMASYNC_POSTWRITE); mcx_dmamem_free(sc, &mxm); free_rq: mcx_dmamem_free(sc, &rx->rx_rq_mem); return (-1); } static int mcx_ready_rq(struct mcx_softc *sc, struct mcx_rx *rx) { struct mcx_cmdq_entry *cqe; struct mcx_dmamem mxm; struct mcx_cmd_modify_rq_in *in; struct mcx_cmd_modify_rq_mb_in *mbin; struct mcx_cmd_modify_rq_out *out; int error; int token; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); token = mcx_cmdq_token(sc); mcx_cmdq_init(sc, cqe, sizeof(*in) + sizeof(*mbin), sizeof(*out), token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_MODIFY_RQ); in->cmd_op_mod = htobe16(0); in->cmd_rq_state = htobe32((MCX_QUEUE_STATE_RST << 28) | rx->rx_rqn); if (mcx_cmdq_mboxes_alloc(sc, &mxm, 1, &cqe->cq_input_ptr, token) != 0) { printf("%s: unable to allocate modify rq mailbox\n", DEVNAME(sc)); return (-1); } mbin = mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0)); mbin->cmd_rq_ctx.rq_flags = htobe32( MCX_QUEUE_STATE_RDY << MCX_RQ_CTX_STATE_SHIFT); mcx_cmdq_mboxes_sign(&mxm, 1); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: modify rq timeout\n", DEVNAME(sc)); goto free; } if (mcx_cmdq_verify(cqe) != 0) { printf("%s: modify rq command corrupt\n", DEVNAME(sc)); goto free; } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf("%s: modify rq failed (%x, %x)\n", DEVNAME(sc), out->cmd_status, be32toh(out->cmd_syndrome)); error = -1; goto free; } free: mcx_dmamem_free(sc, &mxm); return (error); } static int mcx_destroy_rq(struct mcx_softc *sc, struct mcx_rx *rx) { struct mcx_cmdq_entry *cqe; struct mcx_cmd_destroy_rq_in *in; struct mcx_cmd_destroy_rq_out *out; int error; int token; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); token = mcx_cmdq_token(sc); mcx_cmdq_init(sc, cqe, sizeof(*in), sizeof(*out), token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_DESTROY_RQ); in->cmd_op_mod = htobe16(0); in->cmd_rqn = htobe32(rx->rx_rqn); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: destroy rq timeout\n", DEVNAME(sc)); return error; } if (mcx_cmdq_verify(cqe) != 0) { printf("%s: destroy rq command corrupt\n", DEVNAME(sc)); return error; } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf("%s: destroy rq failed (%x, %x)\n", DEVNAME(sc), out->cmd_status, be32toh(out->cmd_syndrome)); return -1; } bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&sc->sc_doorbell_mem), rx->rx_doorbell, sizeof(uint32_t), BUS_DMASYNC_POSTWRITE); bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&rx->rx_rq_mem), 0, MCX_DMA_LEN(&rx->rx_rq_mem), BUS_DMASYNC_POSTWRITE); mcx_dmamem_free(sc, &rx->rx_rq_mem); rx->rx_rqn = 0; return 0; } static int mcx_create_tir_direct(struct mcx_softc *sc, struct mcx_rx *rx, int *tirn) { struct mcx_cmdq_entry *cqe; struct mcx_dmamem mxm; struct mcx_cmd_create_tir_in *in; struct mcx_cmd_create_tir_mb_in *mbin; struct mcx_cmd_create_tir_out *out; int error; int token; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); token = mcx_cmdq_token(sc); mcx_cmdq_init(sc, cqe, sizeof(*in) + sizeof(*mbin), sizeof(*out), token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_CREATE_TIR); in->cmd_op_mod = htobe16(0); if (mcx_cmdq_mboxes_alloc(sc, &mxm, 1, &cqe->cq_input_ptr, token) != 0) { printf("%s: unable to allocate create tir mailbox\n", DEVNAME(sc)); return (-1); } mbin = mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0)); /* leave disp_type = 0, so packets get sent to the inline rqn */ mbin->cmd_inline_rqn = htobe32(rx->rx_rqn); mbin->cmd_tdomain = htobe32(sc->sc_tdomain); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: create tir timeout\n", DEVNAME(sc)); goto free; } if (mcx_cmdq_verify(cqe) != 0) { printf("%s: create tir command corrupt\n", DEVNAME(sc)); goto free; } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf("%s: create tir failed (%x, %x)\n", DEVNAME(sc), out->cmd_status, be32toh(out->cmd_syndrome)); error = -1; goto free; } *tirn = mcx_get_id(out->cmd_tirn); free: mcx_dmamem_free(sc, &mxm); return (error); } static int mcx_create_tir_indirect(struct mcx_softc *sc, int rqtn, uint32_t hash_sel, int *tirn) { struct mcx_cmdq_entry *cqe; struct mcx_dmamem mxm; struct mcx_cmd_create_tir_in *in; struct mcx_cmd_create_tir_mb_in *mbin; struct mcx_cmd_create_tir_out *out; int error; int token; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); token = mcx_cmdq_token(sc); mcx_cmdq_init(sc, cqe, sizeof(*in) + sizeof(*mbin), sizeof(*out), token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_CREATE_TIR); in->cmd_op_mod = htobe16(0); if (mcx_cmdq_mboxes_alloc(sc, &mxm, 1, &cqe->cq_input_ptr, token) != 0) { printf("%s: unable to allocate create tir mailbox\n", DEVNAME(sc)); return (-1); } mbin = mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0)); mbin->cmd_disp_type = htobe32(MCX_TIR_CTX_DISP_TYPE_INDIRECT << MCX_TIR_CTX_DISP_TYPE_SHIFT); mbin->cmd_indir_table = htobe32(rqtn); mbin->cmd_tdomain = htobe32(sc->sc_tdomain | MCX_TIR_CTX_HASH_TOEPLITZ << MCX_TIR_CTX_HASH_SHIFT); mbin->cmd_rx_hash_sel_outer = htobe32(hash_sel); stoeplitz_to_key(&mbin->cmd_rx_hash_key, sizeof(mbin->cmd_rx_hash_key)); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: create tir timeout\n", DEVNAME(sc)); goto free; } if (mcx_cmdq_verify(cqe) != 0) { printf("%s: create tir command corrupt\n", DEVNAME(sc)); goto free; } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf("%s: create tir failed (%x, %x)\n", DEVNAME(sc), out->cmd_status, be32toh(out->cmd_syndrome)); error = -1; goto free; } *tirn = mcx_get_id(out->cmd_tirn); free: mcx_dmamem_free(sc, &mxm); return (error); } static int mcx_destroy_tir(struct mcx_softc *sc, int tirn) { struct mcx_cmdq_entry *cqe; struct mcx_cmd_destroy_tir_in *in; struct mcx_cmd_destroy_tir_out *out; int error; int token; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); token = mcx_cmdq_token(sc); mcx_cmdq_init(sc, cqe, sizeof(*in), sizeof(*out), token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_DESTROY_TIR); in->cmd_op_mod = htobe16(0); in->cmd_tirn = htobe32(tirn); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: destroy tir timeout\n", DEVNAME(sc)); return error; } if (mcx_cmdq_verify(cqe) != 0) { printf("%s: destroy tir command corrupt\n", DEVNAME(sc)); return error; } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf("%s: destroy tir failed (%x, %x)\n", DEVNAME(sc), out->cmd_status, be32toh(out->cmd_syndrome)); return -1; } return (0); } static int mcx_create_sq(struct mcx_softc *sc, struct mcx_tx *tx, int uar, int db, int cqn) { struct mcx_cmdq_entry *cqe; struct mcx_dmamem mxm; struct mcx_cmd_create_sq_in *in; struct mcx_sq_ctx *mbin; struct mcx_cmd_create_sq_out *out; int error; uint64_t *pas; int insize, npages, paslen, token; tx->tx_doorbell = MCX_WQ_DOORBELL_BASE + (db * MCX_WQ_DOORBELL_STRIDE) + 4; npages = howmany((1 << MCX_LOG_SQ_SIZE) * sizeof(struct mcx_sq_entry), MCX_PAGE_SIZE); paslen = npages * sizeof(*pas); insize = sizeof(struct mcx_sq_ctx) + paslen; if (mcx_dmamem_alloc(sc, &tx->tx_sq_mem, npages * MCX_PAGE_SIZE, MCX_PAGE_SIZE) != 0) { printf("%s: unable to allocate send queue memory\n", DEVNAME(sc)); return (-1); } cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); token = mcx_cmdq_token(sc); mcx_cmdq_init(sc, cqe, sizeof(*in) + insize + paslen, sizeof(*out), token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_CREATE_SQ); in->cmd_op_mod = htobe16(0); if (mcx_cmdq_mboxes_alloc(sc, &mxm, howmany(insize, MCX_CMDQ_MAILBOX_DATASIZE), &cqe->cq_input_ptr, token) != 0) { printf("%s: unable to allocate create sq mailboxen\n", DEVNAME(sc)); goto free_sq; } mbin = (struct mcx_sq_ctx *) (((char *)mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0))) + 0x10); mbin->sq_flags = htobe32(MCX_SQ_CTX_RLKEY | (1 << MCX_SQ_CTX_MIN_WQE_INLINE_SHIFT)); mbin->sq_cqn = htobe32(cqn); mbin->sq_tis_lst_sz = htobe32(1 << MCX_SQ_CTX_TIS_LST_SZ_SHIFT); mbin->sq_tis_num = htobe32(sc->sc_tis); mbin->sq_wq.wq_type = MCX_WQ_CTX_TYPE_CYCLIC; mbin->sq_wq.wq_pd = htobe32(sc->sc_pd); mbin->sq_wq.wq_uar_page = htobe32(uar); mbin->sq_wq.wq_doorbell = htobe64(MCX_DMA_DVA(&sc->sc_doorbell_mem) + tx->tx_doorbell); mbin->sq_wq.wq_log_stride = htobe16(MCX_LOG_SQ_ENTRY_SIZE); mbin->sq_wq.wq_log_size = MCX_LOG_SQ_SIZE; bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&tx->tx_sq_mem), 0, MCX_DMA_LEN(&tx->tx_sq_mem), BUS_DMASYNC_PREWRITE); /* physical addresses follow the mailbox in data */ mcx_cmdq_mboxes_pas(&mxm, sizeof(*mbin) + 0x10, npages, &tx->tx_sq_mem); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: create sq timeout\n", DEVNAME(sc)); goto free_mxm; } if (mcx_cmdq_verify(cqe) != 0) { printf("%s: create sq command corrupt\n", DEVNAME(sc)); goto free_mxm; } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf("%s: create sq failed (%x, %x)\n", DEVNAME(sc), out->cmd_status, be32toh(out->cmd_syndrome)); goto free_mxm; } tx->tx_uar = uar; tx->tx_sqn = mcx_get_id(out->cmd_sqn); mcx_dmamem_free(sc, &mxm); bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&sc->sc_doorbell_mem), tx->tx_doorbell, sizeof(uint32_t), BUS_DMASYNC_PREWRITE); return (0); free_mxm: bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&tx->tx_sq_mem), 0, MCX_DMA_LEN(&tx->tx_sq_mem), BUS_DMASYNC_POSTWRITE); mcx_dmamem_free(sc, &mxm); free_sq: mcx_dmamem_free(sc, &tx->tx_sq_mem); return (-1); } static int mcx_destroy_sq(struct mcx_softc *sc, struct mcx_tx *tx) { struct mcx_cmdq_entry *cqe; struct mcx_cmd_destroy_sq_in *in; struct mcx_cmd_destroy_sq_out *out; int error; int token; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); token = mcx_cmdq_token(sc); mcx_cmdq_init(sc, cqe, sizeof(*in), sizeof(*out), token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_DESTROY_SQ); in->cmd_op_mod = htobe16(0); in->cmd_sqn = htobe32(tx->tx_sqn); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: destroy sq timeout\n", DEVNAME(sc)); return error; } if (mcx_cmdq_verify(cqe) != 0) { printf("%s: destroy sq command corrupt\n", DEVNAME(sc)); return error; } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf("%s: destroy sq failed (%x, %x)\n", DEVNAME(sc), out->cmd_status, be32toh(out->cmd_syndrome)); return -1; } bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&sc->sc_doorbell_mem), tx->tx_doorbell, sizeof(uint32_t), BUS_DMASYNC_POSTWRITE); bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&tx->tx_sq_mem), 0, MCX_DMA_LEN(&tx->tx_sq_mem), BUS_DMASYNC_POSTWRITE); mcx_dmamem_free(sc, &tx->tx_sq_mem); tx->tx_sqn = 0; return 0; } static int mcx_ready_sq(struct mcx_softc *sc, struct mcx_tx *tx) { struct mcx_cmdq_entry *cqe; struct mcx_dmamem mxm; struct mcx_cmd_modify_sq_in *in; struct mcx_cmd_modify_sq_mb_in *mbin; struct mcx_cmd_modify_sq_out *out; int error; int token; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); token = mcx_cmdq_token(sc); mcx_cmdq_init(sc, cqe, sizeof(*in) + sizeof(*mbin), sizeof(*out), token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_MODIFY_SQ); in->cmd_op_mod = htobe16(0); in->cmd_sq_state = htobe32((MCX_QUEUE_STATE_RST << 28) | tx->tx_sqn); if (mcx_cmdq_mboxes_alloc(sc, &mxm, 1, &cqe->cq_input_ptr, token) != 0) { printf("%s: unable to allocate modify sq mailbox\n", DEVNAME(sc)); return (-1); } mbin = mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0)); mbin->cmd_sq_ctx.sq_flags = htobe32( MCX_QUEUE_STATE_RDY << MCX_SQ_CTX_STATE_SHIFT); mcx_cmdq_mboxes_sign(&mxm, 1); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: modify sq timeout\n", DEVNAME(sc)); goto free; } if (mcx_cmdq_verify(cqe) != 0) { printf("%s: modify sq command corrupt\n", DEVNAME(sc)); goto free; } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf("%s: modify sq failed (%x, %x)\n", DEVNAME(sc), out->cmd_status, be32toh(out->cmd_syndrome)); error = -1; goto free; } free: mcx_dmamem_free(sc, &mxm); return (error); } static int mcx_create_tis(struct mcx_softc *sc, int *tis) { struct mcx_cmdq_entry *cqe; struct mcx_dmamem mxm; struct mcx_cmd_create_tis_in *in; struct mcx_cmd_create_tis_mb_in *mbin; struct mcx_cmd_create_tis_out *out; int error; int token; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); token = mcx_cmdq_token(sc); mcx_cmdq_init(sc, cqe, sizeof(*in) + sizeof(*mbin), sizeof(*out), token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_CREATE_TIS); in->cmd_op_mod = htobe16(0); if (mcx_cmdq_mboxes_alloc(sc, &mxm, 1, &cqe->cq_input_ptr, token) != 0) { printf("%s: unable to allocate create tis mailbox\n", DEVNAME(sc)); return (-1); } mbin = mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0)); mbin->cmd_tdomain = htobe32(sc->sc_tdomain); mcx_cmdq_mboxes_sign(&mxm, 1); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: create tis timeout\n", DEVNAME(sc)); goto free; } if (mcx_cmdq_verify(cqe) != 0) { printf("%s: create tis command corrupt\n", DEVNAME(sc)); goto free; } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf("%s: create tis failed (%x, %x)\n", DEVNAME(sc), out->cmd_status, be32toh(out->cmd_syndrome)); error = -1; goto free; } *tis = mcx_get_id(out->cmd_tisn); free: mcx_dmamem_free(sc, &mxm); return (error); } static int mcx_destroy_tis(struct mcx_softc *sc, int tis) { struct mcx_cmdq_entry *cqe; struct mcx_cmd_destroy_tis_in *in; struct mcx_cmd_destroy_tis_out *out; int error; int token; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); token = mcx_cmdq_token(sc); mcx_cmdq_init(sc, cqe, sizeof(*in), sizeof(*out), token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_DESTROY_TIS); in->cmd_op_mod = htobe16(0); in->cmd_tisn = htobe32(tis); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: destroy tis timeout\n", DEVNAME(sc)); return error; } if (mcx_cmdq_verify(cqe) != 0) { printf("%s: destroy tis command corrupt\n", DEVNAME(sc)); return error; } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf("%s: destroy tis failed (%x, %x)\n", DEVNAME(sc), out->cmd_status, be32toh(out->cmd_syndrome)); return -1; } return 0; } static int mcx_create_rqt(struct mcx_softc *sc, int size, int *rqns, int *rqt) { struct mcx_cmdq_entry *cqe; struct mcx_dmamem mxm; struct mcx_cmd_create_rqt_in *in; struct mcx_cmd_create_rqt_mb_in *mbin; struct mcx_cmd_create_rqt_out *out; struct mcx_rqt_ctx *rqt_ctx; int *rqtn; int error; int token; int i; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); token = mcx_cmdq_token(sc); mcx_cmdq_init(sc, cqe, sizeof(*in) + sizeof(*mbin) + (size * sizeof(int)), sizeof(*out), token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_CREATE_RQT); in->cmd_op_mod = htobe16(0); if (mcx_cmdq_mboxes_alloc(sc, &mxm, 1, &cqe->cq_input_ptr, token) != 0) { printf("%s: unable to allocate create rqt mailbox\n", DEVNAME(sc)); return (-1); } mbin = mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0)); rqt_ctx = &mbin->cmd_rqt; rqt_ctx->cmd_rqt_max_size = htobe16(sc->sc_max_rqt_size); rqt_ctx->cmd_rqt_actual_size = htobe16(size); /* rqt list follows the rqt context */ rqtn = (int *)(rqt_ctx + 1); for (i = 0; i < size; i++) { rqtn[i] = htobe32(rqns[i]); } mcx_cmdq_mboxes_sign(&mxm, 1); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: create rqt timeout\n", DEVNAME(sc)); goto free; } if (mcx_cmdq_verify(cqe) != 0) { printf("%s: create rqt command corrupt\n", DEVNAME(sc)); goto free; } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf("%s: create rqt failed (%x, %x)\n", DEVNAME(sc), out->cmd_status, be32toh(out->cmd_syndrome)); error = -1; goto free; } *rqt = mcx_get_id(out->cmd_rqtn); return (0); free: mcx_dmamem_free(sc, &mxm); return (error); } static int mcx_destroy_rqt(struct mcx_softc *sc, int rqt) { struct mcx_cmdq_entry *cqe; struct mcx_cmd_destroy_rqt_in *in; struct mcx_cmd_destroy_rqt_out *out; int error; int token; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); token = mcx_cmdq_token(sc); mcx_cmdq_init(sc, cqe, sizeof(*in), sizeof(*out), token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_DESTROY_RQT); in->cmd_op_mod = htobe16(0); in->cmd_rqtn = htobe32(rqt); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: destroy rqt timeout\n", DEVNAME(sc)); return error; } if (mcx_cmdq_verify(cqe) != 0) { printf("%s: destroy rqt command corrupt\n", DEVNAME(sc)); return error; } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf("%s: destroy rqt failed (%x, %x)\n", DEVNAME(sc), out->cmd_status, be32toh(out->cmd_syndrome)); return -1; } return 0; } #if 0 static int mcx_alloc_flow_counter(struct mcx_softc *sc, int i) { struct mcx_cmdq_entry *cqe; struct mcx_cmd_alloc_flow_counter_in *in; struct mcx_cmd_alloc_flow_counter_out *out; int error; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); mcx_cmdq_init(sc, cqe, sizeof(*in), sizeof(*out), mcx_cmdq_token(sc)); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_ALLOC_FLOW_COUNTER); in->cmd_op_mod = htobe16(0); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: alloc flow counter timeout\n", DEVNAME(sc)); return (-1); } if (mcx_cmdq_verify(cqe) != 0) { printf("%s: alloc flow counter command corrupt\n", DEVNAME(sc)); return (-1); } out = (struct mcx_cmd_alloc_flow_counter_out *)cqe->cq_output_data; if (out->cmd_status != MCX_CQ_STATUS_OK) { printf("%s: alloc flow counter failed (%x)\n", DEVNAME(sc), out->cmd_status); return (-1); } sc->sc_flow_counter_id[i] = be16toh(out->cmd_flow_counter_id); printf("flow counter id %d = %d\n", i, sc->sc_flow_counter_id[i]); return (0); } #endif static int mcx_create_flow_table(struct mcx_softc *sc, int log_size, int level, int *flow_table_id) { struct mcx_cmdq_entry *cqe; struct mcx_dmamem mxm; struct mcx_cmd_create_flow_table_in *in; struct mcx_cmd_create_flow_table_mb_in *mbin; struct mcx_cmd_create_flow_table_out *out; int error; int token; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); token = mcx_cmdq_token(sc); mcx_cmdq_init(sc, cqe, sizeof(*in) + sizeof(*mbin), sizeof(*out), token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_CREATE_FLOW_TABLE); in->cmd_op_mod = htobe16(0); if (mcx_cmdq_mboxes_alloc(sc, &mxm, 1, &cqe->cq_input_ptr, token) != 0) { printf("%s: unable to allocate create flow table mailbox\n", DEVNAME(sc)); return (-1); } mbin = mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0)); mbin->cmd_table_type = MCX_FLOW_TABLE_TYPE_RX; mbin->cmd_ctx.ft_log_size = log_size; mbin->cmd_ctx.ft_level = level; mcx_cmdq_mboxes_sign(&mxm, 1); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: create flow table timeout\n", DEVNAME(sc)); goto free; } if (mcx_cmdq_verify(cqe) != 0) { printf("%s: create flow table command corrupt\n", DEVNAME(sc)); goto free; } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf("%s: create flow table failed (%x, %x)\n", DEVNAME(sc), out->cmd_status, be32toh(out->cmd_syndrome)); error = -1; goto free; } *flow_table_id = mcx_get_id(out->cmd_table_id); free: mcx_dmamem_free(sc, &mxm); return (error); } static int mcx_set_flow_table_root(struct mcx_softc *sc, int flow_table_id) { struct mcx_cmdq_entry *cqe; struct mcx_dmamem mxm; struct mcx_cmd_set_flow_table_root_in *in; struct mcx_cmd_set_flow_table_root_mb_in *mbin; struct mcx_cmd_set_flow_table_root_out *out; int error; int token; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); token = mcx_cmdq_token(sc); mcx_cmdq_init(sc, cqe, sizeof(*in) + sizeof(*mbin), sizeof(*out), token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_SET_FLOW_TABLE_ROOT); in->cmd_op_mod = htobe16(0); if (mcx_cmdq_mboxes_alloc(sc, &mxm, 1, &cqe->cq_input_ptr, token) != 0) { printf("%s: unable to allocate set flow table root mailbox\n", DEVNAME(sc)); return (-1); } mbin = mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0)); mbin->cmd_table_type = MCX_FLOW_TABLE_TYPE_RX; mbin->cmd_table_id = htobe32(flow_table_id); mcx_cmdq_mboxes_sign(&mxm, 1); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: set flow table root timeout\n", DEVNAME(sc)); goto free; } if (mcx_cmdq_verify(cqe) != 0) { printf("%s: set flow table root command corrupt\n", DEVNAME(sc)); goto free; } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf("%s: set flow table root failed (%x, %x)\n", DEVNAME(sc), out->cmd_status, be32toh(out->cmd_syndrome)); error = -1; goto free; } free: mcx_dmamem_free(sc, &mxm); return (error); } static int mcx_destroy_flow_table(struct mcx_softc *sc, int flow_table_id) { struct mcx_cmdq_entry *cqe; struct mcx_dmamem mxm; struct mcx_cmd_destroy_flow_table_in *in; struct mcx_cmd_destroy_flow_table_mb_in *mb; struct mcx_cmd_destroy_flow_table_out *out; int error; int token; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); token = mcx_cmdq_token(sc); mcx_cmdq_init(sc, cqe, sizeof(*in) + sizeof(*mb), sizeof(*out), token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_DESTROY_FLOW_TABLE); in->cmd_op_mod = htobe16(0); if (mcx_cmdq_mboxes_alloc(sc, &mxm, 1, &cqe->cq_input_ptr, token) != 0) { printf("%s: unable to allocate destroy flow table mailbox\n", DEVNAME(sc)); return (-1); } mb = mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0)); mb->cmd_table_type = MCX_FLOW_TABLE_TYPE_RX; mb->cmd_table_id = htobe32(flow_table_id); mcx_cmdq_mboxes_sign(&mxm, 1); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: destroy flow table timeout\n", DEVNAME(sc)); goto free; } if (mcx_cmdq_verify(cqe) != 0) { printf("%s: destroy flow table command corrupt\n", DEVNAME(sc)); goto free; } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf("%s: destroy flow table failed (%x, %x)\n", DEVNAME(sc), out->cmd_status, be32toh(out->cmd_syndrome)); error = -1; goto free; } free: mcx_dmamem_free(sc, &mxm); return (error); } static int mcx_create_flow_group(struct mcx_softc *sc, int flow_table_id, int group, int start, int size, int match_enable, struct mcx_flow_match *match) { struct mcx_cmdq_entry *cqe; struct mcx_dmamem mxm; struct mcx_cmd_create_flow_group_in *in; struct mcx_cmd_create_flow_group_mb_in *mbin; struct mcx_cmd_create_flow_group_out *out; struct mcx_flow_group *mfg; int error; int token; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); token = mcx_cmdq_token(sc); mcx_cmdq_init(sc, cqe, sizeof(*in) + sizeof(*mbin), sizeof(*out), token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_CREATE_FLOW_GROUP); in->cmd_op_mod = htobe16(0); if (mcx_cmdq_mboxes_alloc(sc, &mxm, 2, &cqe->cq_input_ptr, token) != 0) { printf("%s: unable to allocate create flow group mailbox\n", DEVNAME(sc)); return (-1); } mbin = mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0)); mbin->cmd_table_type = MCX_FLOW_TABLE_TYPE_RX; mbin->cmd_table_id = htobe32(flow_table_id); mbin->cmd_start_flow_index = htobe32(start); mbin->cmd_end_flow_index = htobe32(start + (size - 1)); mbin->cmd_match_criteria_enable = match_enable; memcpy(&mbin->cmd_match_criteria, match, sizeof(*match)); mcx_cmdq_mboxes_sign(&mxm, 2); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: create flow group timeout\n", DEVNAME(sc)); goto free; } if (mcx_cmdq_verify(cqe) != 0) { printf("%s: create flow group command corrupt\n", DEVNAME(sc)); goto free; } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf("%s: create flow group failed (%x, %x)\n", DEVNAME(sc), out->cmd_status, be32toh(out->cmd_syndrome)); error = -1; goto free; } mfg = &sc->sc_flow_group[group]; mfg->g_id = mcx_get_id(out->cmd_group_id); mfg->g_table = flow_table_id; mfg->g_start = start; mfg->g_size = size; free: mcx_dmamem_free(sc, &mxm); return (error); } static int mcx_destroy_flow_group(struct mcx_softc *sc, int group) { struct mcx_cmdq_entry *cqe; struct mcx_dmamem mxm; struct mcx_cmd_destroy_flow_group_in *in; struct mcx_cmd_destroy_flow_group_mb_in *mb; struct mcx_cmd_destroy_flow_group_out *out; struct mcx_flow_group *mfg; int error; int token; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); token = mcx_cmdq_token(sc); mcx_cmdq_init(sc, cqe, sizeof(*in) + sizeof(*mb), sizeof(*out), token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_DESTROY_FLOW_GROUP); in->cmd_op_mod = htobe16(0); if (mcx_cmdq_mboxes_alloc(sc, &mxm, 2, &cqe->cq_input_ptr, token) != 0) { printf("%s: unable to allocate destroy flow group mailbox\n", DEVNAME(sc)); return (-1); } mb = mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0)); mb->cmd_table_type = MCX_FLOW_TABLE_TYPE_RX; mfg = &sc->sc_flow_group[group]; mb->cmd_table_id = htobe32(mfg->g_table); mb->cmd_group_id = htobe32(mfg->g_id); mcx_cmdq_mboxes_sign(&mxm, 2); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: destroy flow group timeout\n", DEVNAME(sc)); goto free; } if (mcx_cmdq_verify(cqe) != 0) { printf("%s: destroy flow group command corrupt\n", DEVNAME(sc)); goto free; } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf("%s: destroy flow group failed (%x, %x)\n", DEVNAME(sc), out->cmd_status, be32toh(out->cmd_syndrome)); error = -1; goto free; } mfg->g_id = -1; mfg->g_table = -1; mfg->g_size = 0; mfg->g_start = 0; free: mcx_dmamem_free(sc, &mxm); return (error); } static int mcx_set_flow_table_entry_mac(struct mcx_softc *sc, int group, int index, const uint8_t *macaddr, uint32_t dest) { struct mcx_cmdq_entry *cqe; struct mcx_dmamem mxm; struct mcx_cmd_set_flow_table_entry_in *in; struct mcx_cmd_set_flow_table_entry_mb_in *mbin; struct mcx_cmd_set_flow_table_entry_out *out; struct mcx_flow_group *mfg; uint32_t *pdest; int error; int token; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); token = mcx_cmdq_token(sc); mcx_cmdq_init(sc, cqe, sizeof(*in) + sizeof(*mbin) + sizeof(*pdest), sizeof(*out), token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_SET_FLOW_TABLE_ENTRY); in->cmd_op_mod = htobe16(0); if (mcx_cmdq_mboxes_alloc(sc, &mxm, 2, &cqe->cq_input_ptr, token) != 0) { printf("%s: unable to allocate set flow table entry mailbox\n", DEVNAME(sc)); return (-1); } mbin = mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0)); mbin->cmd_table_type = MCX_FLOW_TABLE_TYPE_RX; mfg = &sc->sc_flow_group[group]; mbin->cmd_table_id = htobe32(mfg->g_table); mbin->cmd_flow_index = htobe32(mfg->g_start + index); mbin->cmd_flow_ctx.fc_group_id = htobe32(mfg->g_id); /* flow context ends at offset 0x330, 0x130 into the second mbox */ pdest = (uint32_t *) (((char *)mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 1))) + 0x130); mbin->cmd_flow_ctx.fc_action = htobe32(MCX_FLOW_CONTEXT_ACTION_FORWARD); mbin->cmd_flow_ctx.fc_dest_list_size = htobe32(1); *pdest = htobe32(dest); /* the only thing we match on at the moment is the dest mac address */ if (macaddr != NULL) { memcpy(mbin->cmd_flow_ctx.fc_match_value.mc_dest_mac, macaddr, ETHER_ADDR_LEN); } mcx_cmdq_mboxes_sign(&mxm, 2); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: set flow table entry timeout\n", DEVNAME(sc)); goto free; } if (mcx_cmdq_verify(cqe) != 0) { printf("%s: set flow table entry command corrupt\n", DEVNAME(sc)); goto free; } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf("%s: set flow table entry failed (%x, %x)\n", DEVNAME(sc), out->cmd_status, be32toh(out->cmd_syndrome)); error = -1; goto free; } free: mcx_dmamem_free(sc, &mxm); return (error); } static int mcx_set_flow_table_entry_proto(struct mcx_softc *sc, int group, int index, int ethertype, int ip_proto, uint32_t dest) { struct mcx_cmdq_entry *cqe; struct mcx_dmamem mxm; struct mcx_cmd_set_flow_table_entry_in *in; struct mcx_cmd_set_flow_table_entry_mb_in *mbin; struct mcx_cmd_set_flow_table_entry_out *out; struct mcx_flow_group *mfg; uint32_t *pdest; int error; int token; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); token = mcx_cmdq_token(sc); mcx_cmdq_init(sc, cqe, sizeof(*in) + sizeof(*mbin) + sizeof(*pdest), sizeof(*out), token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_SET_FLOW_TABLE_ENTRY); in->cmd_op_mod = htobe16(0); if (mcx_cmdq_mboxes_alloc(sc, &mxm, 2, &cqe->cq_input_ptr, token) != 0) { printf("%s: unable to allocate set flow table entry mailbox\n", DEVNAME(sc)); return (-1); } mbin = mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0)); mbin->cmd_table_type = MCX_FLOW_TABLE_TYPE_RX; mfg = &sc->sc_flow_group[group]; mbin->cmd_table_id = htobe32(mfg->g_table); mbin->cmd_flow_index = htobe32(mfg->g_start + index); mbin->cmd_flow_ctx.fc_group_id = htobe32(mfg->g_id); /* flow context ends at offset 0x330, 0x130 into the second mbox */ pdest = (uint32_t *) (((char *)mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 1))) + 0x130); mbin->cmd_flow_ctx.fc_action = htobe32(MCX_FLOW_CONTEXT_ACTION_FORWARD); mbin->cmd_flow_ctx.fc_dest_list_size = htobe32(1); *pdest = htobe32(dest); mbin->cmd_flow_ctx.fc_match_value.mc_ethertype = htobe16(ethertype); mbin->cmd_flow_ctx.fc_match_value.mc_ip_proto = ip_proto; mcx_cmdq_mboxes_sign(&mxm, 2); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: set flow table entry timeout\n", DEVNAME(sc)); goto free; } if (mcx_cmdq_verify(cqe) != 0) { printf("%s: set flow table entry command corrupt\n", DEVNAME(sc)); goto free; } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf("%s: set flow table entry failed (%x, %x)\n", DEVNAME(sc), out->cmd_status, be32toh(out->cmd_syndrome)); error = -1; goto free; } free: mcx_dmamem_free(sc, &mxm); return (error); } static int mcx_delete_flow_table_entry(struct mcx_softc *sc, int group, int index) { struct mcx_cmdq_entry *cqe; struct mcx_dmamem mxm; struct mcx_cmd_delete_flow_table_entry_in *in; struct mcx_cmd_delete_flow_table_entry_mb_in *mbin; struct mcx_cmd_delete_flow_table_entry_out *out; struct mcx_flow_group *mfg; int error; int token; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); token = mcx_cmdq_token(sc); mcx_cmdq_init(sc, cqe, sizeof(*in) + sizeof(*mbin), sizeof(*out), token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_DELETE_FLOW_TABLE_ENTRY); in->cmd_op_mod = htobe16(0); if (mcx_cmdq_mboxes_alloc(sc, &mxm, 2, &cqe->cq_input_ptr, token) != 0) { printf("%s: unable to allocate " "delete flow table entry mailbox\n", DEVNAME(sc)); return (-1); } mbin = mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0)); mbin->cmd_table_type = MCX_FLOW_TABLE_TYPE_RX; mfg = &sc->sc_flow_group[group]; mbin->cmd_table_id = htobe32(mfg->g_table); mbin->cmd_flow_index = htobe32(mfg->g_start + index); mcx_cmdq_mboxes_sign(&mxm, 2); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: delete flow table entry timeout\n", DEVNAME(sc)); goto free; } if (mcx_cmdq_verify(cqe) != 0) { printf("%s: delete flow table entry command corrupt\n", DEVNAME(sc)); goto free; } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf("%s: delete flow table entry %d:%d failed (%x, %x)\n", DEVNAME(sc), group, index, out->cmd_status, be32toh(out->cmd_syndrome)); error = -1; goto free; } free: mcx_dmamem_free(sc, &mxm); return (error); } #if 0 int mcx_dump_flow_table(struct mcx_softc *sc, int flow_table_id) { struct mcx_dmamem mxm; struct mcx_cmdq_entry *cqe; struct mcx_cmd_query_flow_table_in *in; struct mcx_cmd_query_flow_table_mb_in *mbin; struct mcx_cmd_query_flow_table_out *out; struct mcx_cmd_query_flow_table_mb_out *mbout; uint8_t token = mcx_cmdq_token(sc); int error; int i; uint8_t *dump; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); mcx_cmdq_init(sc, cqe, sizeof(*in) + sizeof(*mbin), sizeof(*out) + sizeof(*mbout) + 16, token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_QUERY_FLOW_TABLE); in->cmd_op_mod = htobe16(0); CTASSERT(sizeof(*mbin) <= MCX_CMDQ_MAILBOX_DATASIZE); CTASSERT(sizeof(*mbout) <= MCX_CMDQ_MAILBOX_DATASIZE); if (mcx_cmdq_mboxes_alloc(sc, &mxm, 2, &cqe->cq_output_ptr, token) != 0) { printf(", unable to allocate query flow table mailboxes\n"); return (-1); } cqe->cq_input_ptr = cqe->cq_output_ptr; mbin = mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0)); mbin->cmd_table_type = 0; mbin->cmd_table_id = htobe32(flow_table_id); mcx_cmdq_mboxes_sign(&mxm, 1); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: query flow table timeout\n", DEVNAME(sc)); goto free; } error = mcx_cmdq_verify(cqe); if (error != 0) { printf("%s: query flow table reply corrupt\n", DEVNAME(sc)); goto free; } out = mcx_cmdq_out(cqe); switch (out->cmd_status) { case MCX_CQ_STATUS_OK: break; default: printf("%s: query flow table failed (%x/%x)\n", DEVNAME(sc), out->cmd_status, be32toh(out->cmd_syndrome)); error = -1; goto free; } mbout = (struct mcx_cmd_query_flow_table_mb_out *) (mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0))); dump = (uint8_t *)mbout + 8; for (i = 0; i < sizeof(struct mcx_flow_table_ctx); i++) { printf("%.2x ", dump[i]); if (i % 16 == 15) printf("\n"); } free: mcx_cq_mboxes_free(sc, &mxm); return (error); } int mcx_dump_flow_table_entry(struct mcx_softc *sc, int flow_table_id, int index) { struct mcx_dmamem mxm; struct mcx_cmdq_entry *cqe; struct mcx_cmd_query_flow_table_entry_in *in; struct mcx_cmd_query_flow_table_entry_mb_in *mbin; struct mcx_cmd_query_flow_table_entry_out *out; struct mcx_cmd_query_flow_table_entry_mb_out *mbout; uint8_t token = mcx_cmdq_token(sc); int error; int i; uint8_t *dump; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); mcx_cmdq_init(sc, cqe, sizeof(*in) + sizeof(*mbin), sizeof(*out) + sizeof(*mbout) + 16, token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_QUERY_FLOW_TABLE_ENTRY); in->cmd_op_mod = htobe16(0); CTASSERT(sizeof(*mbin) <= MCX_CMDQ_MAILBOX_DATASIZE); CTASSERT(sizeof(*mbout) <= MCX_CMDQ_MAILBOX_DATASIZE*2); if (mcx_cmdq_mboxes_alloc(sc, &mxm, 2, &cqe->cq_output_ptr, token) != 0) { printf(", unable to allocate " "query flow table entry mailboxes\n"); return (-1); } cqe->cq_input_ptr = cqe->cq_output_ptr; mbin = mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0)); mbin->cmd_table_type = 0; mbin->cmd_table_id = htobe32(flow_table_id); mbin->cmd_flow_index = htobe32(index); mcx_cmdq_mboxes_sign(&mxm, 1); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: query flow table entry timeout\n", DEVNAME(sc)); goto free; } error = mcx_cmdq_verify(cqe); if (error != 0) { printf("%s: query flow table entry reply corrupt\n", DEVNAME(sc)); goto free; } out = mcx_cmdq_out(cqe); switch (out->cmd_status) { case MCX_CQ_STATUS_OK: break; default: printf("%s: query flow table entry failed (%x/%x)\n", DEVNAME(sc), out->cmd_status, be32toh(out->cmd_syndrome)); error = -1; goto free; } mbout = (struct mcx_cmd_query_flow_table_entry_mb_out *) (mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0))); dump = (uint8_t *)mbout; for (i = 0; i < MCX_CMDQ_MAILBOX_DATASIZE; i++) { printf("%.2x ", dump[i]); if (i % 16 == 15) printf("\n"); } free: mcx_cq_mboxes_free(sc, &mxm); return (error); } int mcx_dump_flow_group(struct mcx_softc *sc, int flow_table_id) { struct mcx_dmamem mxm; struct mcx_cmdq_entry *cqe; struct mcx_cmd_query_flow_group_in *in; struct mcx_cmd_query_flow_group_mb_in *mbin; struct mcx_cmd_query_flow_group_out *out; struct mcx_cmd_query_flow_group_mb_out *mbout; uint8_t token = mcx_cmdq_token(sc); int error; int i; uint8_t *dump; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); mcx_cmdq_init(sc, cqe, sizeof(*in) + sizeof(*mbin), sizeof(*out) + sizeof(*mbout) + 16, token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_QUERY_FLOW_GROUP); in->cmd_op_mod = htobe16(0); CTASSERT(sizeof(*mbin) <= MCX_CMDQ_MAILBOX_DATASIZE); CTASSERT(sizeof(*mbout) <= MCX_CMDQ_MAILBOX_DATASIZE*2); if (mcx_cmdq_mboxes_alloc(sc, &mxm, 2, &cqe->cq_output_ptr, token) != 0) { printf(", unable to allocate query flow group mailboxes\n"); return (-1); } cqe->cq_input_ptr = cqe->cq_output_ptr; mbin = mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0)); mbin->cmd_table_type = 0; mbin->cmd_table_id = htobe32(flow_table_id); mbin->cmd_group_id = htobe32(sc->sc_flow_group_id); mcx_cmdq_mboxes_sign(&mxm, 1); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: query flow group timeout\n", DEVNAME(sc)); goto free; } error = mcx_cmdq_verify(cqe); if (error != 0) { printf("%s: query flow group reply corrupt\n", DEVNAME(sc)); goto free; } out = mcx_cmdq_out(cqe); switch (out->cmd_status) { case MCX_CQ_STATUS_OK: break; default: printf("%s: query flow group failed (%x/%x)\n", DEVNAME(sc), out->cmd_status, be32toh(out->cmd_syndrome)); error = -1; goto free; } mbout = (struct mcx_cmd_query_flow_group_mb_out *) (mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0))); dump = (uint8_t *)mbout; for (i = 0; i < MCX_CMDQ_MAILBOX_DATASIZE; i++) { printf("%.2x ", dump[i]); if (i % 16 == 15) printf("\n"); } dump = (uint8_t *)(mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 1))); for (i = 0; i < MCX_CMDQ_MAILBOX_DATASIZE; i++) { printf("%.2x ", dump[i]); if (i % 16 == 15) printf("\n"); } free: mcx_cq_mboxes_free(sc, &mxm); return (error); } static int mcx_dump_counters(struct mcx_softc *sc) { struct mcx_dmamem mxm; struct mcx_cmdq_entry *cqe; struct mcx_cmd_query_vport_counters_in *in; struct mcx_cmd_query_vport_counters_mb_in *mbin; struct mcx_cmd_query_vport_counters_out *out; struct mcx_nic_vport_counters *counters; int error, token; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); token = mcx_cmdq_token(sc); mcx_cmdq_init(sc, cqe, sizeof(*in) + sizeof(*mbin), sizeof(*out) + sizeof(*counters), token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_QUERY_VPORT_COUNTERS); in->cmd_op_mod = htobe16(0); if (mcx_cmdq_mboxes_alloc(sc, &mxm, 1, &cqe->cq_output_ptr, token) != 0) { printf(", unable to allocate " "query nic vport counters mailboxen\n"); return (-1); } cqe->cq_input_ptr = cqe->cq_output_ptr; mbin = mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0)); mbin->cmd_clear = 0x80; mcx_cmdq_mboxes_sign(&mxm, 1); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: query nic vport counters timeout\n", DEVNAME(sc)); goto free; } if (mcx_cmdq_verify(cqe) != 0) { printf("%s: query nic vport counters command corrupt\n", DEVNAME(sc)); goto free; } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf("%s: query nic vport counters failed (%x, %x)\n", DEVNAME(sc), out->cmd_status, betoh32(out->cmd_syndrome)); error = -1; goto free; } counters = (struct mcx_nic_vport_counters *) (mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0))); if (counters->rx_bcast.packets + counters->tx_bcast.packets + counters->rx_ucast.packets + counters->tx_ucast.packets + counters->rx_err.packets + counters->tx_err.packets) printf("%s: err %llx/%llx uc %llx/%llx bc %llx/%llx\n", DEVNAME(sc), betoh64(counters->tx_err.packets), betoh64(counters->rx_err.packets), betoh64(counters->tx_ucast.packets), betoh64(counters->rx_ucast.packets), betoh64(counters->tx_bcast.packets), betoh64(counters->rx_bcast.packets)); free: mcx_dmamem_free(sc, &mxm); return (error); } static int mcx_dump_flow_counter(struct mcx_softc *sc, int index, const char *what) { struct mcx_dmamem mxm; struct mcx_cmdq_entry *cqe; struct mcx_cmd_query_flow_counter_in *in; struct mcx_cmd_query_flow_counter_mb_in *mbin; struct mcx_cmd_query_flow_counter_out *out; struct mcx_counter *counters; int error, token; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); token = mcx_cmdq_token(sc); mcx_cmdq_init(sc, cqe, sizeof(*in) + sizeof(*mbin), sizeof(*out) + sizeof(*counters), token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_QUERY_FLOW_COUNTER); in->cmd_op_mod = htobe16(0); if (mcx_cmdq_mboxes_alloc(sc, &mxm, 1, &cqe->cq_output_ptr, token) != 0) { printf(", unable to allocate query flow counter mailboxen\n"); return (-1); } cqe->cq_input_ptr = cqe->cq_output_ptr; mbin = mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0)); mbin->cmd_flow_counter_id = htobe16(sc->sc_flow_counter_id[index]); mbin->cmd_clear = 0x80; mcx_cmdq_mboxes_sign(&mxm, 1); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: query flow counter timeout\n", DEVNAME(sc)); goto free; } if (mcx_cmdq_verify(cqe) != 0) { printf("%s: query flow counter command corrupt\n", DEVNAME(sc)); goto free; } out = mcx_cmdq_out(cqe); if (out->cmd_status != MCX_CQ_STATUS_OK) { printf("%s: query flow counter failed (%x, %x)\n", DEVNAME(sc), out->cmd_status, betoh32(out->cmd_syndrome)); error = -1; goto free; } counters = (struct mcx_counter *) (mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0))); if (counters->packets) printf("%s: %s inflow %llx\n", DEVNAME(sc), what, betoh64(counters->packets)); free: mcx_dmamem_free(sc, &mxm); return (error); } #endif #if NKSTAT > 0 int mcx_query_rq(struct mcx_softc *sc, struct mcx_rx *rx, struct mcx_rq_ctx *rq_ctx) { struct mcx_dmamem mxm; struct mcx_cmdq_entry *cqe; struct mcx_cmd_query_rq_in *in; struct mcx_cmd_query_rq_out *out; struct mcx_cmd_query_rq_mb_out *mbout; uint8_t token = mcx_cmdq_token(sc); int error; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); mcx_cmdq_init(sc, cqe, sizeof(*in), sizeof(*out) + sizeof(*mbout) + 16, token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_QUERY_RQ); in->cmd_op_mod = htobe16(0); in->cmd_rqn = htobe32(rx->rx_rqn); CTASSERT(sizeof(*mbout) <= MCX_CMDQ_MAILBOX_DATASIZE*2); if (mcx_cmdq_mboxes_alloc(sc, &mxm, 2, &cqe->cq_output_ptr, token) != 0) { printf("%s: unable to allocate query rq mailboxes\n", DEVNAME(sc)); return (-1); } mcx_cmdq_mboxes_sign(&mxm, 1); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: query rq timeout\n", DEVNAME(sc)); goto free; } error = mcx_cmdq_verify(cqe); if (error != 0) { printf("%s: query rq reply corrupt\n", DEVNAME(sc)); goto free; } out = mcx_cmdq_out(cqe); switch (out->cmd_status) { case MCX_CQ_STATUS_OK: break; default: printf("%s: query rq failed (%x/%x)\n", DEVNAME(sc), out->cmd_status, be32toh(out->cmd_syndrome)); error = -1; goto free; } mbout = (struct mcx_cmd_query_rq_mb_out *) (mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0))); memcpy(rq_ctx, &mbout->cmd_ctx, sizeof(*rq_ctx)); free: mcx_cq_mboxes_free(sc, &mxm); return (error); } int mcx_query_sq(struct mcx_softc *sc, struct mcx_tx *tx, struct mcx_sq_ctx *sq_ctx) { struct mcx_dmamem mxm; struct mcx_cmdq_entry *cqe; struct mcx_cmd_query_sq_in *in; struct mcx_cmd_query_sq_out *out; struct mcx_cmd_query_sq_mb_out *mbout; uint8_t token = mcx_cmdq_token(sc); int error; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); mcx_cmdq_init(sc, cqe, sizeof(*in), sizeof(*out) + sizeof(*mbout) + 16, token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_QUERY_SQ); in->cmd_op_mod = htobe16(0); in->cmd_sqn = htobe32(tx->tx_sqn); CTASSERT(sizeof(*mbout) <= MCX_CMDQ_MAILBOX_DATASIZE*2); if (mcx_cmdq_mboxes_alloc(sc, &mxm, 2, &cqe->cq_output_ptr, token) != 0) { printf("%s: unable to allocate query sq mailboxes\n", DEVNAME(sc)); return (-1); } mcx_cmdq_mboxes_sign(&mxm, 1); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: query sq timeout\n", DEVNAME(sc)); goto free; } error = mcx_cmdq_verify(cqe); if (error != 0) { printf("%s: query sq reply corrupt\n", DEVNAME(sc)); goto free; } out = mcx_cmdq_out(cqe); switch (out->cmd_status) { case MCX_CQ_STATUS_OK: break; default: printf("%s: query sq failed (%x/%x)\n", DEVNAME(sc), out->cmd_status, be32toh(out->cmd_syndrome)); error = -1; goto free; } mbout = (struct mcx_cmd_query_sq_mb_out *) (mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0))); memcpy(sq_ctx, &mbout->cmd_ctx, sizeof(*sq_ctx)); free: mcx_cq_mboxes_free(sc, &mxm); return (error); } int mcx_query_cq(struct mcx_softc *sc, struct mcx_cq *cq, struct mcx_cq_ctx *cq_ctx) { struct mcx_dmamem mxm; struct mcx_cmdq_entry *cqe; struct mcx_cmd_query_cq_in *in; struct mcx_cmd_query_cq_out *out; struct mcx_cq_ctx *ctx; uint8_t token = mcx_cmdq_token(sc); int error; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); mcx_cmdq_init(sc, cqe, sizeof(*in), sizeof(*out) + sizeof(*ctx) + 16, token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_QUERY_CQ); in->cmd_op_mod = htobe16(0); in->cmd_cqn = htobe32(cq->cq_n); CTASSERT(sizeof(*ctx) <= MCX_CMDQ_MAILBOX_DATASIZE*2); if (mcx_cmdq_mboxes_alloc(sc, &mxm, 2, &cqe->cq_output_ptr, token) != 0) { printf("%s: unable to allocate query cq mailboxes\n", DEVNAME(sc)); return (-1); } mcx_cmdq_mboxes_sign(&mxm, 1); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: query cq timeout\n", DEVNAME(sc)); goto free; } if (mcx_cmdq_verify(cqe) != 0) { printf("%s: query cq reply corrupt\n", DEVNAME(sc)); goto free; } out = mcx_cmdq_out(cqe); switch (out->cmd_status) { case MCX_CQ_STATUS_OK: break; default: printf("%s: query qc failed (%x/%x)\n", DEVNAME(sc), out->cmd_status, be32toh(out->cmd_syndrome)); error = -1; goto free; } ctx = (struct mcx_cq_ctx *)(mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0))); memcpy(cq_ctx, ctx, sizeof(*cq_ctx)); free: mcx_dmamem_free(sc, &mxm); return (error); } int mcx_query_eq(struct mcx_softc *sc, struct mcx_eq *eq, struct mcx_eq_ctx *eq_ctx) { struct mcx_dmamem mxm; struct mcx_cmdq_entry *cqe; struct mcx_cmd_query_eq_in *in; struct mcx_cmd_query_eq_out *out; struct mcq_eq_ctx *ctx; uint8_t token = mcx_cmdq_token(sc); int error; cqe = MCX_DMA_KVA(&sc->sc_cmdq_mem); mcx_cmdq_init(sc, cqe, sizeof(*in), sizeof(*out) + sizeof(*ctx) + 16, token); in = mcx_cmdq_in(cqe); in->cmd_opcode = htobe16(MCX_CMD_QUERY_EQ); in->cmd_op_mod = htobe16(0); in->cmd_eqn = htobe32(eq->eq_n); CTASSERT(sizeof(*ctx) <= MCX_CMDQ_MAILBOX_DATASIZE*2); if (mcx_cmdq_mboxes_alloc(sc, &mxm, 2, &cqe->cq_output_ptr, token) != 0) { printf("%s: unable to allocate query eq mailboxes\n", DEVNAME(sc)); return (-1); } mcx_cmdq_mboxes_sign(&mxm, 1); mcx_cmdq_post(sc, cqe, 0); error = mcx_cmdq_poll(sc, cqe, 1000); if (error != 0) { printf("%s: query eq timeout\n", DEVNAME(sc)); goto free; } if (mcx_cmdq_verify(cqe) != 0) { printf("%s: query eq reply corrupt\n", DEVNAME(sc)); goto free; } out = mcx_cmdq_out(cqe); switch (out->cmd_status) { case MCX_CQ_STATUS_OK: break; default: printf("%s: query eq failed (%x/%x)\n", DEVNAME(sc), out->cmd_status, be32toh(out->cmd_syndrome)); error = -1; goto free; } ctx = (struct mcx_eq_ctx *)(mcx_cq_mbox_data(mcx_cq_mbox(&mxm, 0))); memcpy(eq_ctx, ctx, sizeof(*eq_ctx)); free: mcx_dmamem_free(sc, &mxm); return (error); } #endif /* NKSTAT > 0 */ static inline unsigned int mcx_rx_fill_slots(struct mcx_softc *sc, struct mcx_rx *rx, uint nslots) { struct mcx_rq_entry *ring, *rqe; struct mcx_slot *ms; struct mbuf *m; uint slot, p, fills; ring = MCX_DMA_KVA(&rx->rx_rq_mem); p = rx->rx_prod; bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&rx->rx_rq_mem), 0, MCX_DMA_LEN(&rx->rx_rq_mem), BUS_DMASYNC_POSTWRITE); slot = (p % (1 << MCX_LOG_RQ_SIZE)); rqe = ring; for (fills = 0; fills < nslots; fills++) { slot = p % (1 << MCX_LOG_RQ_SIZE); ms = &rx->rx_slots[slot]; rqe = &ring[slot]; m = NULL; MGETHDR(m, M_DONTWAIT, MT_DATA); if (m == NULL) break; MCLGET(m, M_DONTWAIT); if ((m->m_flags & M_EXT) == 0) { m_freem(m); break; } m->m_len = m->m_pkthdr.len = sc->sc_hardmtu; m_adj(m, m->m_ext.ext_size - sc->sc_rxbufsz); m_adj(m, ETHER_ALIGN); if (bus_dmamap_load_mbuf(sc->sc_dmat, ms->ms_map, m, BUS_DMA_NOWAIT) != 0) { m_freem(m); break; } bus_dmamap_sync(sc->sc_dmat, ms->ms_map, 0, ms->ms_map->dm_mapsize, BUS_DMASYNC_PREREAD); ms->ms_m = m; be32enc(&rqe->rqe_byte_count, ms->ms_map->dm_segs[0].ds_len); be64enc(&rqe->rqe_addr, ms->ms_map->dm_segs[0].ds_addr); be32enc(&rqe->rqe_lkey, sc->sc_lkey); p++; } bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&rx->rx_rq_mem), 0, MCX_DMA_LEN(&rx->rx_rq_mem), BUS_DMASYNC_PREWRITE); rx->rx_prod = p; bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&sc->sc_doorbell_mem), rx->rx_doorbell, sizeof(uint32_t), BUS_DMASYNC_POSTWRITE); be32enc(MCX_DMA_OFF(&sc->sc_doorbell_mem, rx->rx_doorbell), p & MCX_WQ_DOORBELL_MASK); bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&sc->sc_doorbell_mem), rx->rx_doorbell, sizeof(uint32_t), BUS_DMASYNC_PREWRITE); return (nslots - fills); } static int mcx_rx_fill(struct mcx_softc *sc, struct mcx_rx *rx) { u_int slots; slots = mcx_rxr_get(&rx->rx_rxr, (1 << MCX_LOG_RQ_SIZE)); if (slots == 0) return (1); slots = mcx_rx_fill_slots(sc, rx, slots); mcx_rxr_put(&rx->rx_rxr, slots); return (0); } void mcx_refill(void *xrx) { struct mcx_rx *rx = xrx; struct mcx_softc *sc = rx->rx_softc; mcx_rx_fill(sc, rx); if (mcx_rxr_inuse(&rx->rx_rxr) == 0) callout_schedule(&rx->rx_refill, 1); } static int mcx_process_txeof(struct mcx_softc *sc, struct mcx_tx *tx, struct mcx_cq_entry *cqe) { struct mcx_slot *ms; bus_dmamap_t map; int slot, slots; slot = be16toh(cqe->cq_wqe_count) % (1 << MCX_LOG_SQ_SIZE); ms = &tx->tx_slots[slot]; map = ms->ms_map; bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize, BUS_DMASYNC_POSTWRITE); slots = 1; if (map->dm_nsegs > 1) slots += (map->dm_nsegs+2) / MCX_SQ_SEGS_PER_SLOT; bus_dmamap_unload(sc->sc_dmat, map); m_freem(ms->ms_m); ms->ms_m = NULL; return (slots); } static uint64_t mcx_uptime(void) { struct timespec ts; nanouptime(&ts); return ((uint64_t)ts.tv_sec * 1000000000 + (uint64_t)ts.tv_nsec); } static void mcx_calibrate_first(struct mcx_softc *sc) { struct mcx_calibration *c = &sc->sc_calibration[0]; int s; sc->sc_calibration_gen = 0; s = splhigh(); /* crit_enter? */ c->c_ubase = mcx_uptime(); c->c_tbase = mcx_timer(sc); splx(s); c->c_ratio = 0; #if notyet callout_schedule(&sc->sc_calibrate, MCX_CALIBRATE_FIRST * hz); #endif } #define MCX_TIMESTAMP_SHIFT 24 static void mcx_calibrate(void *arg) { struct mcx_softc *sc = arg; struct mcx_calibration *nc, *pc; uint64_t udiff, tdiff; unsigned int gen; int s; if (!ISSET(sc->sc_ec.ec_if.if_flags, IFF_RUNNING)) return; callout_schedule(&sc->sc_calibrate, MCX_CALIBRATE_NORMAL * hz); gen = sc->sc_calibration_gen; pc = &sc->sc_calibration[gen % __arraycount(sc->sc_calibration)]; gen++; nc = &sc->sc_calibration[gen % __arraycount(sc->sc_calibration)]; nc->c_uptime = pc->c_ubase; nc->c_timestamp = pc->c_tbase; s = splhigh(); /* crit_enter? */ nc->c_ubase = mcx_uptime(); nc->c_tbase = mcx_timer(sc); splx(s); udiff = nc->c_ubase - nc->c_uptime; tdiff = nc->c_tbase - nc->c_timestamp; /* * udiff is the wall clock time between calibration ticks, * which should be 32 seconds or 32 billion nanoseconds. if * we squint, 1 billion nanoseconds is kind of like a 32 bit * number, so 32 billion should still have a lot of high bits * spare. we use this space by shifting the nanoseconds up * 24 bits so we have a nice big number to divide by the * number of mcx timer ticks. */ nc->c_ratio = (udiff << MCX_TIMESTAMP_SHIFT) / tdiff; membar_producer(); sc->sc_calibration_gen = gen; } static int mcx_process_rx(struct mcx_softc *sc, struct mcx_rx *rx, struct mcx_cq_entry *cqe, struct mcx_mbufq *mq, const struct mcx_calibration *c) { struct ifnet *ifp = &sc->sc_ec.ec_if; struct mcx_slot *ms; struct mbuf *m; uint32_t flags, len; int slot; len = be32dec(&cqe->cq_byte_cnt); slot = be16toh(cqe->cq_wqe_count) % (1 << MCX_LOG_RQ_SIZE); ms = &rx->rx_slots[slot]; bus_dmamap_sync(sc->sc_dmat, ms->ms_map, 0, len, BUS_DMASYNC_POSTREAD); bus_dmamap_unload(sc->sc_dmat, ms->ms_map); m = ms->ms_m; ms->ms_m = NULL; m_set_rcvif(m, &sc->sc_ec.ec_if); m->m_pkthdr.len = m->m_len = len; #if 0 if (cqe->cq_rx_hash_type) { m->m_pkthdr.ph_flowid = be32toh(cqe->cq_rx_hash); m->m_pkthdr.csum_flags |= M_FLOWID; } #endif flags = be32dec(&cqe->cq_flags); if (flags & MCX_CQ_ENTRY_FLAGS_L3_OK) { if (ifp->if_capenable & IFCAP_CSUM_IPv4_Rx) m->m_pkthdr.csum_flags |= M_CSUM_IPv4; } if (flags & MCX_CQ_ENTRY_FLAGS_L4_OK) { if (ifp->if_capenable & IFCAP_CSUM_TCPv4_Rx) m->m_pkthdr.csum_flags |= M_CSUM_TCPv4; if (ifp->if_capenable & IFCAP_CSUM_TCPv6_Rx) m->m_pkthdr.csum_flags |= M_CSUM_TCPv6; if (ifp->if_capenable & IFCAP_CSUM_UDPv4_Rx) m->m_pkthdr.csum_flags |= M_CSUM_UDPv4; if (ifp->if_capenable & IFCAP_CSUM_UDPv6_Rx) m->m_pkthdr.csum_flags |= M_CSUM_UDPv6; } if (flags & MCX_CQ_ENTRY_FLAGS_CV) { vlan_set_tag(m, flags & MCX_CQ_ENTRY_FLAGS_VLAN_MASK); } #if notyet if (ISSET(sc->sc_ec.ec_if.if_flags, IFF_LINK0) && c->c_ratio) { uint64_t t = be64dec(&cqe->cq_timestamp); t -= c->c_timestamp; t *= c->c_ratio; t >>= MCX_TIMESTAMP_SHIFT; t += c->c_uptime; m->m_pkthdr.ph_timestamp = t; SET(m->m_pkthdr.csum_flags, M_TIMESTAMP); } #endif MBUFQ_ENQUEUE(mq, m); return (1); } static struct mcx_cq_entry * mcx_next_cq_entry(struct mcx_softc *sc, struct mcx_cq *cq) { struct mcx_cq_entry *cqe; int next; cqe = (struct mcx_cq_entry *)MCX_DMA_KVA(&cq->cq_mem); next = cq->cq_cons % (1 << MCX_LOG_CQ_SIZE); if ((cqe[next].cq_opcode_owner & MCX_CQ_ENTRY_FLAG_OWNER) == ((cq->cq_cons >> MCX_LOG_CQ_SIZE) & 1)) { return (&cqe[next]); } return (NULL); } static void mcx_arm_cq(struct mcx_softc *sc, struct mcx_cq *cq, int uar) { struct mcx_cq_doorbell *db; bus_size_t offset; uint32_t val; uint64_t uval; val = ((cq->cq_count) & 3) << MCX_CQ_DOORBELL_ARM_CMD_SN_SHIFT; val |= (cq->cq_cons & MCX_CQ_DOORBELL_ARM_CI_MASK); db = MCX_DMA_OFF(&sc->sc_doorbell_mem, cq->cq_doorbell); bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&sc->sc_doorbell_mem), cq->cq_doorbell, sizeof(*db), BUS_DMASYNC_POSTWRITE); be32enc(&db->db_update_ci, cq->cq_cons & MCX_CQ_DOORBELL_ARM_CI_MASK); be32enc(&db->db_arm_ci, val); bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&sc->sc_doorbell_mem), cq->cq_doorbell, sizeof(*db), BUS_DMASYNC_PREWRITE); offset = (MCX_PAGE_SIZE * uar) + MCX_UAR_CQ_DOORBELL; uval = (uint64_t)val << 32; uval |= cq->cq_n; bus_space_write_8(sc->sc_memt, sc->sc_memh, offset, htobe64(uval)); mcx_bar(sc, offset, sizeof(uval), BUS_SPACE_BARRIER_WRITE); } void mcx_process_cq(struct mcx_softc *sc, struct mcx_queues *q, struct mcx_cq *cq) { struct mcx_rx *rx = &q->q_rx; struct mcx_tx *tx = &q->q_tx; struct ifnet *ifp = &sc->sc_ec.ec_if; const struct mcx_calibration *c; unsigned int gen; struct mcx_cq_entry *cqe; struct mcx_mbufq mq; struct mbuf *m; int rxfree, txfree; MBUFQ_INIT(&mq); gen = sc->sc_calibration_gen; membar_consumer(); c = &sc->sc_calibration[gen % __arraycount(sc->sc_calibration)]; bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&cq->cq_mem), 0, MCX_DMA_LEN(&cq->cq_mem), BUS_DMASYNC_POSTREAD); rxfree = 0; txfree = 0; while ((cqe = mcx_next_cq_entry(sc, cq))) { uint8_t opcode; opcode = (cqe->cq_opcode_owner >> MCX_CQ_ENTRY_OPCODE_SHIFT); switch (opcode) { case MCX_CQ_ENTRY_OPCODE_REQ: txfree += mcx_process_txeof(sc, tx, cqe); break; case MCX_CQ_ENTRY_OPCODE_SEND: rxfree += mcx_process_rx(sc, rx, cqe, &mq, c); break; case MCX_CQ_ENTRY_OPCODE_REQ_ERR: case MCX_CQ_ENTRY_OPCODE_SEND_ERR: /* uint8_t *cqp = (uint8_t *)cqe; */ /* printf("%s: cq completion error: %x\n", DEVNAME(sc), cqp[0x37]); */ break; default: /* printf("%s: cq completion opcode %x??\n", DEVNAME(sc), opcode); */ break; } cq->cq_cons++; } bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&cq->cq_mem), 0, MCX_DMA_LEN(&cq->cq_mem), BUS_DMASYNC_PREREAD); if (rxfree > 0) { mcx_rxr_put(&rx->rx_rxr, rxfree); while (MBUFQ_FIRST(&mq) != NULL) { MBUFQ_DEQUEUE(&mq, m); if_percpuq_enqueue(ifp->if_percpuq, m); } mcx_rx_fill(sc, rx); if (mcx_rxr_inuse(&rx->rx_rxr) == 0) callout_schedule(&rx->rx_refill, 1); } cq->cq_count++; mcx_arm_cq(sc, cq, q->q_uar); if (txfree > 0) { tx->tx_cons += txfree; if_schedule_deferred_start(ifp); } } static void mcx_arm_eq(struct mcx_softc *sc, struct mcx_eq *eq, int uar) { bus_size_t offset; uint32_t val; offset = (MCX_PAGE_SIZE * uar) + MCX_UAR_EQ_DOORBELL_ARM; val = (eq->eq_n << 24) | (eq->eq_cons & 0xffffff); mcx_wr(sc, offset, val); mcx_bar(sc, offset, sizeof(val), BUS_SPACE_BARRIER_WRITE); } static struct mcx_eq_entry * mcx_next_eq_entry(struct mcx_softc *sc, struct mcx_eq *eq) { struct mcx_eq_entry *eqe; int next; eqe = (struct mcx_eq_entry *)MCX_DMA_KVA(&eq->eq_mem); next = eq->eq_cons % (1 << MCX_LOG_EQ_SIZE); if ((eqe[next].eq_owner & 1) == ((eq->eq_cons >> MCX_LOG_EQ_SIZE) & 1)) { eq->eq_cons++; return (&eqe[next]); } return (NULL); } int mcx_admin_intr(void *xsc) { struct mcx_softc *sc = (struct mcx_softc *)xsc; struct mcx_eq *eq = &sc->sc_admin_eq; struct mcx_eq_entry *eqe; bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&eq->eq_mem), 0, MCX_DMA_LEN(&eq->eq_mem), BUS_DMASYNC_POSTREAD); while ((eqe = mcx_next_eq_entry(sc, eq)) != NULL) { switch (eqe->eq_event_type) { case MCX_EVENT_TYPE_LAST_WQE: /* printf("%s: last wqe reached?\n", DEVNAME(sc)); */ break; case MCX_EVENT_TYPE_CQ_ERROR: /* printf("%s: cq error\n", DEVNAME(sc)); */ break; case MCX_EVENT_TYPE_CMD_COMPLETION: /* wakeup probably */ break; case MCX_EVENT_TYPE_PORT_CHANGE: workqueue_enqueue(sc->sc_workq, &sc->sc_port_change, NULL); break; default: /* printf("%s: something happened\n", DEVNAME(sc)); */ break; } } bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&eq->eq_mem), 0, MCX_DMA_LEN(&eq->eq_mem), BUS_DMASYNC_PREREAD); mcx_arm_eq(sc, eq, sc->sc_uar); return (1); } int mcx_cq_intr(void *xq) { struct mcx_queues *q = (struct mcx_queues *)xq; struct mcx_softc *sc = q->q_sc; struct mcx_eq *eq = &q->q_eq; struct mcx_eq_entry *eqe; int cqn; bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&eq->eq_mem), 0, MCX_DMA_LEN(&eq->eq_mem), BUS_DMASYNC_POSTREAD); while ((eqe = mcx_next_eq_entry(sc, eq)) != NULL) { switch (eqe->eq_event_type) { case MCX_EVENT_TYPE_COMPLETION: cqn = be32toh(eqe->eq_event_data[6]); if (cqn == q->q_cq.cq_n) mcx_process_cq(sc, q, &q->q_cq); break; } } bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&eq->eq_mem), 0, MCX_DMA_LEN(&eq->eq_mem), BUS_DMASYNC_PREREAD); mcx_arm_eq(sc, eq, q->q_uar); return (1); } static void mcx_free_slots(struct mcx_softc *sc, struct mcx_slot *slots, int allocated, int total) { struct mcx_slot *ms; int i = allocated; while (i-- > 0) { ms = &slots[i]; bus_dmamap_destroy(sc->sc_dmat, ms->ms_map); if (ms->ms_m != NULL) m_freem(ms->ms_m); } kmem_free(slots, total * sizeof(*ms)); } static int mcx_queue_up(struct mcx_softc *sc, struct mcx_queues *q) { struct mcx_rx *rx; struct mcx_tx *tx; struct mcx_slot *ms; int i; rx = &q->q_rx; rx->rx_slots = kmem_zalloc(sizeof(*ms) * (1 << MCX_LOG_RQ_SIZE), KM_SLEEP); for (i = 0; i < (1 << MCX_LOG_RQ_SIZE); i++) { ms = &rx->rx_slots[i]; if (bus_dmamap_create(sc->sc_dmat, sc->sc_hardmtu, 1, sc->sc_hardmtu, 0, BUS_DMA_WAITOK | BUS_DMA_ALLOCNOW, &ms->ms_map) != 0) { printf("%s: failed to allocate rx dma maps\n", DEVNAME(sc)); goto destroy_rx_slots; } } tx = &q->q_tx; tx->tx_slots = kmem_zalloc(sizeof(*ms) * (1 << MCX_LOG_SQ_SIZE), KM_SLEEP); for (i = 0; i < (1 << MCX_LOG_SQ_SIZE); i++) { ms = &tx->tx_slots[i]; if (bus_dmamap_create(sc->sc_dmat, sc->sc_hardmtu, MCX_SQ_MAX_SEGMENTS, sc->sc_hardmtu, 0, BUS_DMA_WAITOK | BUS_DMA_ALLOCNOW, &ms->ms_map) != 0) { printf("%s: failed to allocate tx dma maps\n", DEVNAME(sc)); goto destroy_tx_slots; } } if (mcx_create_cq(sc, &q->q_cq, q->q_uar, q->q_index, q->q_eq.eq_n) != 0) goto destroy_tx_slots; if (mcx_create_sq(sc, tx, q->q_uar, q->q_index, q->q_cq.cq_n) != 0) goto destroy_cq; if (mcx_create_rq(sc, rx, q->q_index, q->q_cq.cq_n) != 0) goto destroy_sq; return 0; destroy_sq: mcx_destroy_sq(sc, tx); destroy_cq: mcx_destroy_cq(sc, &q->q_cq); destroy_tx_slots: mcx_free_slots(sc, tx->tx_slots, i, (1 << MCX_LOG_SQ_SIZE)); tx->tx_slots = NULL; i = (1 << MCX_LOG_RQ_SIZE); destroy_rx_slots: mcx_free_slots(sc, rx->rx_slots, i, (1 << MCX_LOG_RQ_SIZE)); rx->rx_slots = NULL; return ENOMEM; } static int mcx_rss_group_entry_count(struct mcx_softc *sc, int group) { int i; int count; count = 0; for (i = 0; i < __arraycount(mcx_rss_config); i++) { if (mcx_rss_config[i].flow_group == group) count++; } return count; } static int mcx_init(struct ifnet *ifp) { struct mcx_softc *sc = ifp->if_softc; struct mcx_rx *rx; struct mcx_tx *tx; int i, start, count, flow_group, flow_index; struct mcx_flow_match match_crit; struct mcx_rss_rule *rss; uint32_t dest; int rqns[MCX_MAX_QUEUES] = { 0 }; if (ISSET(ifp->if_flags, IFF_RUNNING)) mcx_stop(ifp, 0); if (mcx_create_tis(sc, &sc->sc_tis) != 0) goto down; for (i = 0; i < sc->sc_nqueues; i++) { if (mcx_queue_up(sc, &sc->sc_queues[i]) != 0) { goto down; } } /* RSS flow table and flow groups */ if (mcx_create_flow_table(sc, MCX_LOG_FLOW_TABLE_SIZE, 1, &sc->sc_rss_flow_table_id) != 0) goto down; dest = MCX_FLOW_CONTEXT_DEST_TYPE_TABLE | sc->sc_rss_flow_table_id; /* L4 RSS flow group (v4/v6 tcp/udp, no fragments) */ memset(&match_crit, 0, sizeof(match_crit)); match_crit.mc_ethertype = 0xffff; match_crit.mc_ip_proto = 0xff; match_crit.mc_vlan_flags = MCX_FLOW_MATCH_IP_FRAG; start = 0; count = mcx_rss_group_entry_count(sc, MCX_FLOW_GROUP_RSS_L4); if (count != 0) { if (mcx_create_flow_group(sc, sc->sc_rss_flow_table_id, MCX_FLOW_GROUP_RSS_L4, start, count, MCX_CREATE_FLOW_GROUP_CRIT_OUTER, &match_crit) != 0) goto down; start += count; } /* L3 RSS flow group (v4/v6, including fragments) */ memset(&match_crit, 0, sizeof(match_crit)); match_crit.mc_ethertype = 0xffff; count = mcx_rss_group_entry_count(sc, MCX_FLOW_GROUP_RSS_L3); if (mcx_create_flow_group(sc, sc->sc_rss_flow_table_id, MCX_FLOW_GROUP_RSS_L3, start, count, MCX_CREATE_FLOW_GROUP_CRIT_OUTER, &match_crit) != 0) goto down; start += count; /* non-RSS flow group */ count = mcx_rss_group_entry_count(sc, MCX_FLOW_GROUP_RSS_NONE); memset(&match_crit, 0, sizeof(match_crit)); if (mcx_create_flow_group(sc, sc->sc_rss_flow_table_id, MCX_FLOW_GROUP_RSS_NONE, start, count, 0, &match_crit) != 0) goto down; /* Root flow table, matching packets based on mac address */ if (mcx_create_flow_table(sc, MCX_LOG_FLOW_TABLE_SIZE, 0, &sc->sc_mac_flow_table_id) != 0) goto down; /* promisc flow group */ start = 0; memset(&match_crit, 0, sizeof(match_crit)); if (mcx_create_flow_group(sc, sc->sc_mac_flow_table_id, MCX_FLOW_GROUP_PROMISC, start, 1, 0, &match_crit) != 0) goto down; sc->sc_promisc_flow_enabled = 0; start++; /* all multicast flow group */ match_crit.mc_dest_mac[0] = 0x01; if (mcx_create_flow_group(sc, sc->sc_mac_flow_table_id, MCX_FLOW_GROUP_ALLMULTI, start, 1, MCX_CREATE_FLOW_GROUP_CRIT_OUTER, &match_crit) != 0) goto down; sc->sc_allmulti_flow_enabled = 0; start++; /* mac address matching flow group */ memset(&match_crit.mc_dest_mac, 0xff, sizeof(match_crit.mc_dest_mac)); if (mcx_create_flow_group(sc, sc->sc_mac_flow_table_id, MCX_FLOW_GROUP_MAC, start, (1 << MCX_LOG_FLOW_TABLE_SIZE) - start, MCX_CREATE_FLOW_GROUP_CRIT_OUTER, &match_crit) != 0) goto down; /* flow table entries for unicast and broadcast */ start = 0; if (mcx_set_flow_table_entry_mac(sc, MCX_FLOW_GROUP_MAC, start, LLADDR(satosdl(ifp->if_dl->ifa_addr)), dest) != 0) goto down; start++; if (mcx_set_flow_table_entry_mac(sc, MCX_FLOW_GROUP_MAC, start, etherbroadcastaddr, dest) != 0) goto down; start++; /* multicast entries go after that */ sc->sc_mcast_flow_base = start; /* re-add any existing multicast flows */ for (i = 0; i < MCX_NUM_MCAST_FLOWS; i++) { if (sc->sc_mcast_flows[i][0] != 0) { mcx_set_flow_table_entry_mac(sc, MCX_FLOW_GROUP_MAC, sc->sc_mcast_flow_base + i, sc->sc_mcast_flows[i], dest); } } if (mcx_set_flow_table_root(sc, sc->sc_mac_flow_table_id) != 0) goto down; /* * the RQT can be any size as long as it's a power of two. * since we also restrict the number of queues to a power of two, * we can just put each rx queue in once. */ for (i = 0; i < sc->sc_nqueues; i++) rqns[i] = sc->sc_queues[i].q_rx.rx_rqn; if (mcx_create_rqt(sc, sc->sc_nqueues, rqns, &sc->sc_rqt) != 0) goto down; start = 0; flow_index = 0; flow_group = -1; for (i = 0; i < __arraycount(mcx_rss_config); i++) { rss = &mcx_rss_config[i]; if (rss->flow_group != flow_group) { flow_group = rss->flow_group; flow_index = 0; } if (rss->hash_sel == 0) { if (mcx_create_tir_direct(sc, &sc->sc_queues[0].q_rx, &sc->sc_tir[i]) != 0) goto down; } else { if (mcx_create_tir_indirect(sc, sc->sc_rqt, rss->hash_sel, &sc->sc_tir[i]) != 0) goto down; } if (mcx_set_flow_table_entry_proto(sc, flow_group, flow_index, rss->ethertype, rss->ip_proto, MCX_FLOW_CONTEXT_DEST_TYPE_TIR | sc->sc_tir[i]) != 0) goto down; flow_index++; } for (i = 0; i < sc->sc_nqueues; i++) { struct mcx_queues *q = &sc->sc_queues[i]; rx = &q->q_rx; tx = &q->q_tx; /* start the queues */ if (mcx_ready_sq(sc, tx) != 0) goto down; if (mcx_ready_rq(sc, rx) != 0) goto down; mcx_rxr_init(&rx->rx_rxr, 1, (1 << MCX_LOG_RQ_SIZE)); rx->rx_prod = 0; mcx_rx_fill(sc, rx); tx->tx_cons = 0; tx->tx_prod = 0; } mcx_calibrate_first(sc); SET(ifp->if_flags, IFF_RUNNING); CLR(ifp->if_flags, IFF_OACTIVE); if_schedule_deferred_start(ifp); return 0; down: mcx_stop(ifp, 0); return EIO; } static void mcx_stop(struct ifnet *ifp, int disable) { struct mcx_softc *sc = ifp->if_softc; struct mcx_rss_rule *rss; int group, i, flow_group, flow_index; CLR(ifp->if_flags, IFF_RUNNING); /* * delete flow table entries first, so no packets can arrive * after the barriers */ if (sc->sc_promisc_flow_enabled) mcx_delete_flow_table_entry(sc, MCX_FLOW_GROUP_PROMISC, 0); if (sc->sc_allmulti_flow_enabled) mcx_delete_flow_table_entry(sc, MCX_FLOW_GROUP_ALLMULTI, 0); mcx_delete_flow_table_entry(sc, MCX_FLOW_GROUP_MAC, 0); mcx_delete_flow_table_entry(sc, MCX_FLOW_GROUP_MAC, 1); for (i = 0; i < MCX_NUM_MCAST_FLOWS; i++) { if (sc->sc_mcast_flows[i][0] != 0) { mcx_delete_flow_table_entry(sc, MCX_FLOW_GROUP_MAC, sc->sc_mcast_flow_base + i); } } flow_group = -1; flow_index = 0; for (i = 0; i < __arraycount(mcx_rss_config); i++) { rss = &mcx_rss_config[i]; if (rss->flow_group != flow_group) { flow_group = rss->flow_group; flow_index = 0; } mcx_delete_flow_table_entry(sc, flow_group, flow_index); mcx_destroy_tir(sc, sc->sc_tir[i]); sc->sc_tir[i] = 0; flow_index++; } for (i = 0; i < sc->sc_nqueues; i++) { callout_halt(&sc->sc_queues[i].q_rx.rx_refill, NULL); } callout_halt(&sc->sc_calibrate, NULL); for (group = 0; group < MCX_NUM_FLOW_GROUPS; group++) { if (sc->sc_flow_group[group].g_id != -1) mcx_destroy_flow_group(sc, group); } if (sc->sc_mac_flow_table_id != -1) { mcx_destroy_flow_table(sc, sc->sc_mac_flow_table_id); sc->sc_mac_flow_table_id = -1; } if (sc->sc_rss_flow_table_id != -1) { mcx_destroy_flow_table(sc, sc->sc_rss_flow_table_id); sc->sc_rss_flow_table_id = -1; } if (sc->sc_rqt != -1) { mcx_destroy_rqt(sc, sc->sc_rqt); sc->sc_rqt = -1; } for (i = 0; i < sc->sc_nqueues; i++) { struct mcx_queues *q = &sc->sc_queues[i]; struct mcx_rx *rx = &q->q_rx; struct mcx_tx *tx = &q->q_tx; struct mcx_cq *cq = &q->q_cq; if (rx->rx_rqn != 0) mcx_destroy_rq(sc, rx); if (tx->tx_sqn != 0) mcx_destroy_sq(sc, tx); if (tx->tx_slots != NULL) { mcx_free_slots(sc, tx->tx_slots, (1 << MCX_LOG_SQ_SIZE), (1 << MCX_LOG_SQ_SIZE)); tx->tx_slots = NULL; } if (rx->rx_slots != NULL) { mcx_free_slots(sc, rx->rx_slots, (1 << MCX_LOG_RQ_SIZE), (1 << MCX_LOG_RQ_SIZE)); rx->rx_slots = NULL; } if (cq->cq_n != 0) mcx_destroy_cq(sc, cq); } if (sc->sc_tis != 0) { mcx_destroy_tis(sc, sc->sc_tis); sc->sc_tis = 0; } } static int mcx_ioctl(struct ifnet *ifp, u_long cmd, void *data) { struct mcx_softc *sc = (struct mcx_softc *)ifp->if_softc; struct ifreq *ifr = (struct ifreq *)data; struct ethercom *ec = &sc->sc_ec; uint8_t addrhi[ETHER_ADDR_LEN], addrlo[ETHER_ADDR_LEN]; struct ether_multi *enm; struct ether_multistep step; int s, i, flags, error = 0; uint32_t dest; s = splnet(); switch (cmd) { case SIOCADDMULTI: if (ether_addmulti(ifreq_getaddr(cmd, ifr), &sc->sc_ec) == ENETRESET) { error = ether_multiaddr(&ifr->ifr_addr, addrlo, addrhi); if (error != 0) { splx(s); return (error); } dest = MCX_FLOW_CONTEXT_DEST_TYPE_TABLE | sc->sc_rss_flow_table_id; for (i = 0; i < MCX_NUM_MCAST_FLOWS; i++) { if (sc->sc_mcast_flows[i][0] == 0) { memcpy(sc->sc_mcast_flows[i], addrlo, ETHER_ADDR_LEN); if (ISSET(ifp->if_flags, IFF_RUNNING)) { mcx_set_flow_table_entry_mac(sc, MCX_FLOW_GROUP_MAC, sc->sc_mcast_flow_base + i, sc->sc_mcast_flows[i], dest); } break; } } if (!ISSET(ifp->if_flags, IFF_ALLMULTI)) { if (i == MCX_NUM_MCAST_FLOWS) { SET(ifp->if_flags, IFF_ALLMULTI); sc->sc_extra_mcast++; error = ENETRESET; } if (memcmp(addrlo, addrhi, ETHER_ADDR_LEN)) { SET(ifp->if_flags, IFF_ALLMULTI); error = ENETRESET; } } } break; case SIOCDELMULTI: if (ether_delmulti(ifreq_getaddr(cmd, ifr), &sc->sc_ec) == ENETRESET) { error = ether_multiaddr(&ifr->ifr_addr, addrlo, addrhi); if (error != 0) { splx(s); return (error); } for (i = 0; i < MCX_NUM_MCAST_FLOWS; i++) { if (memcmp(sc->sc_mcast_flows[i], addrlo, ETHER_ADDR_LEN) == 0) { if (ISSET(ifp->if_flags, IFF_RUNNING)) { mcx_delete_flow_table_entry(sc, MCX_FLOW_GROUP_MAC, sc->sc_mcast_flow_base + i); } sc->sc_mcast_flows[i][0] = 0; break; } } if (i == MCX_NUM_MCAST_FLOWS) sc->sc_extra_mcast--; if (ISSET(ifp->if_flags, IFF_ALLMULTI) && sc->sc_extra_mcast == 0) { flags = 0; ETHER_LOCK(ec); ETHER_FIRST_MULTI(step, ec, enm); while (enm != NULL) { if (memcmp(enm->enm_addrlo, enm->enm_addrhi, ETHER_ADDR_LEN)) { SET(flags, IFF_ALLMULTI); break; } ETHER_NEXT_MULTI(step, enm); } ETHER_UNLOCK(ec); if (!ISSET(flags, IFF_ALLMULTI)) { CLR(ifp->if_flags, IFF_ALLMULTI); error = ENETRESET; } } } break; default: error = ether_ioctl(ifp, cmd, data); } if (error == ENETRESET) { if ((ifp->if_flags & (IFF_UP | IFF_RUNNING)) == (IFF_UP | IFF_RUNNING)) mcx_iff(sc); error = 0; } splx(s); return (error); } #if 0 static int mcx_get_sffpage(struct ifnet *ifp, struct if_sffpage *sff) { struct mcx_softc *sc = (struct mcx_softc *)ifp->if_softc; struct mcx_reg_mcia mcia; struct mcx_reg_pmlp pmlp; int offset, error; /* get module number */ memset(&pmlp, 0, sizeof(pmlp)); pmlp.rp_local_port = 1; error = mcx_access_hca_reg(sc, MCX_REG_PMLP, MCX_REG_OP_READ, &pmlp, sizeof(pmlp)); if (error != 0) { printf("%s: unable to get eeprom module number\n", DEVNAME(sc)); return error; } for (offset = 0; offset < 256; offset += MCX_MCIA_EEPROM_BYTES) { memset(&mcia, 0, sizeof(mcia)); mcia.rm_l = 0; mcia.rm_module = be32toh(pmlp.rp_lane0_mapping) & MCX_PMLP_MODULE_NUM_MASK; mcia.rm_i2c_addr = sff->sff_addr / 2; /* apparently */ mcia.rm_page_num = sff->sff_page; mcia.rm_dev_addr = htobe16(offset); mcia.rm_size = htobe16(MCX_MCIA_EEPROM_BYTES); error = mcx_access_hca_reg(sc, MCX_REG_MCIA, MCX_REG_OP_READ, &mcia, sizeof(mcia)); if (error != 0) { printf("%s: unable to read eeprom at %x\n", DEVNAME(sc), offset); return error; } memcpy(sff->sff_data + offset, mcia.rm_data, MCX_MCIA_EEPROM_BYTES); } return 0; } #endif static int mcx_load_mbuf(struct mcx_softc *sc, struct mcx_slot *ms, struct mbuf *m) { switch (bus_dmamap_load_mbuf(sc->sc_dmat, ms->ms_map, m, BUS_DMA_STREAMING | BUS_DMA_NOWAIT)) { case 0: break; case EFBIG: if (m_defrag(m, M_DONTWAIT) != NULL && bus_dmamap_load_mbuf(sc->sc_dmat, ms->ms_map, m, BUS_DMA_STREAMING | BUS_DMA_NOWAIT) == 0) break; /* FALLTHROUGH */ default: return (1); } ms->ms_m = m; return (0); } static void mcx_send_common_locked(struct ifnet *ifp, struct mcx_tx *tx, bool is_transmit) { struct mcx_softc *sc = ifp->if_softc; struct mcx_sq_entry *sq, *sqe; struct mcx_sq_entry_seg *sqs; struct mcx_slot *ms; bus_dmamap_t map; struct mbuf *m; u_int idx, free, used; uint64_t *bf; uint32_t csum; size_t bf_base; int i, seg, nseg; KASSERT(mutex_owned(&tx->tx_lock)); if ((ifp->if_flags & IFF_RUNNING) == 0) return; bf_base = (tx->tx_uar * MCX_PAGE_SIZE) + MCX_UAR_BF; idx = tx->tx_prod % (1 << MCX_LOG_SQ_SIZE); free = (tx->tx_cons + (1 << MCX_LOG_SQ_SIZE)) - tx->tx_prod; used = 0; bf = NULL; bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&tx->tx_sq_mem), 0, MCX_DMA_LEN(&tx->tx_sq_mem), BUS_DMASYNC_POSTWRITE); sq = (struct mcx_sq_entry *)MCX_DMA_KVA(&tx->tx_sq_mem); for (;;) { if (used + MCX_SQ_ENTRY_MAX_SLOTS >= free) { SET(ifp->if_flags, IFF_OACTIVE); break; } if (is_transmit) { m = pcq_get(tx->tx_pcq); } else { IFQ_DEQUEUE(&ifp->if_snd, m); } if (m == NULL) { break; } sqe = sq + idx; ms = &tx->tx_slots[idx]; memset(sqe, 0, sizeof(*sqe)); /* ctrl segment */ sqe->sqe_opcode_index = htobe32(MCX_SQE_WQE_OPCODE_SEND | ((tx->tx_prod & 0xffff) << MCX_SQE_WQE_INDEX_SHIFT)); /* always generate a completion event */ sqe->sqe_signature = htobe32(MCX_SQE_CE_CQE_ALWAYS); /* eth segment */ csum = 0; if (m->m_pkthdr.csum_flags & M_CSUM_IPv4) csum |= MCX_SQE_L3_CSUM; if (m->m_pkthdr.csum_flags & (M_CSUM_TCPv4 | M_CSUM_UDPv4 | M_CSUM_TCPv6 | M_CSUM_UDPv6)) csum |= MCX_SQE_L4_CSUM; sqe->sqe_mss_csum = htobe32(csum); sqe->sqe_inline_header_size = htobe16(MCX_SQ_INLINE_SIZE); if (vlan_has_tag(m)) { struct ether_vlan_header *evh; evh = (struct ether_vlan_header *) &sqe->sqe_inline_headers; m_copydata(m, 0, ETHER_HDR_LEN, evh); evh->evl_proto = evh->evl_encap_proto; evh->evl_encap_proto = htons(ETHERTYPE_VLAN); evh->evl_tag = htons(vlan_get_tag(m)); m_adj(m, ETHER_HDR_LEN); } else { m_copydata(m, 0, MCX_SQ_INLINE_SIZE, sqe->sqe_inline_headers); m_adj(m, MCX_SQ_INLINE_SIZE); } if (mcx_load_mbuf(sc, ms, m) != 0) { m_freem(m); if_statinc(ifp, if_oerrors); continue; } bf = (uint64_t *)sqe; if (ifp->if_bpf != NULL) bpf_mtap2(ifp->if_bpf, sqe->sqe_inline_headers, MCX_SQ_INLINE_SIZE, m, BPF_D_OUT); map = ms->ms_map; bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize, BUS_DMASYNC_PREWRITE); sqe->sqe_ds_sq_num = htobe32((tx->tx_sqn << MCX_SQE_SQ_NUM_SHIFT) | (map->dm_nsegs + 3)); /* data segment - first wqe has one segment */ sqs = sqe->sqe_segs; seg = 0; nseg = 1; for (i = 0; i < map->dm_nsegs; i++) { if (seg == nseg) { /* next slot */ idx++; if (idx == (1 << MCX_LOG_SQ_SIZE)) idx = 0; tx->tx_prod++; used++; sqs = (struct mcx_sq_entry_seg *)(sq + idx); seg = 0; nseg = MCX_SQ_SEGS_PER_SLOT; } sqs[seg].sqs_byte_count = htobe32(map->dm_segs[i].ds_len); sqs[seg].sqs_lkey = htobe32(sc->sc_lkey); sqs[seg].sqs_addr = htobe64(map->dm_segs[i].ds_addr); seg++; } idx++; if (idx == (1 << MCX_LOG_SQ_SIZE)) idx = 0; tx->tx_prod++; used++; } bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&tx->tx_sq_mem), 0, MCX_DMA_LEN(&tx->tx_sq_mem), BUS_DMASYNC_PREWRITE); if (used) { bus_size_t blueflame; bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&sc->sc_doorbell_mem), tx->tx_doorbell, sizeof(uint32_t), BUS_DMASYNC_POSTWRITE); be32enc(MCX_DMA_OFF(&sc->sc_doorbell_mem, tx->tx_doorbell), tx->tx_prod & MCX_WQ_DOORBELL_MASK); bus_dmamap_sync(sc->sc_dmat, MCX_DMA_MAP(&sc->sc_doorbell_mem), tx->tx_doorbell, sizeof(uint32_t), BUS_DMASYNC_PREWRITE); /* * write the first 64 bits of the last sqe we produced * to the blue flame buffer */ blueflame = bf_base + tx->tx_bf_offset; bus_space_write_8(sc->sc_memt, sc->sc_memh, blueflame, *bf); mcx_bar(sc, blueflame, sizeof(*bf), BUS_SPACE_BARRIER_WRITE); /* next write goes to the other buffer */ tx->tx_bf_offset ^= sc->sc_bf_size; } } static void mcx_start(struct ifnet *ifp) { struct mcx_softc *sc = ifp->if_softc; /* mcx_start() always uses TX ring[0] */ struct mcx_tx *tx = &sc->sc_queues[0].q_tx; mutex_enter(&tx->tx_lock); if (!ISSET(ifp->if_flags, IFF_OACTIVE)) { mcx_send_common_locked(ifp, tx, false); } mutex_exit(&tx->tx_lock); } static int mcx_transmit(struct ifnet *ifp, struct mbuf *m) { struct mcx_softc *sc = ifp->if_softc; struct mcx_tx *tx; tx = &sc->sc_queues[cpu_index(curcpu()) % sc->sc_nqueues].q_tx; if (__predict_false(!pcq_put(tx->tx_pcq, m))) { m_freem(m); return ENOBUFS; } if (mutex_tryenter(&tx->tx_lock)) { mcx_send_common_locked(ifp, tx, true); mutex_exit(&tx->tx_lock); } else { softint_schedule(tx->tx_softint); } return 0; } static void mcx_deferred_transmit(void *arg) { struct mcx_tx *tx = arg; struct mcx_softc *sc = tx->tx_softc; struct ifnet *ifp = &sc->sc_ec.ec_if; mutex_enter(&tx->tx_lock); if (pcq_peek(tx->tx_pcq) != NULL) { mcx_send_common_locked(ifp, tx, true); } mutex_exit(&tx->tx_lock); } static void mcx_media_add_types(struct mcx_softc *sc) { struct mcx_reg_ptys ptys; int i; uint32_t proto_cap; memset(&ptys, 0, sizeof(ptys)); ptys.rp_local_port = 1; ptys.rp_proto_mask = MCX_REG_PTYS_PROTO_MASK_ETH; if (mcx_access_hca_reg(sc, MCX_REG_PTYS, MCX_REG_OP_READ, &ptys, sizeof(ptys)) != 0) { printf("%s: unable to read port type/speed\n", DEVNAME(sc)); return; } proto_cap = be32toh(ptys.rp_eth_proto_cap); for (i = 0; i < __arraycount(mcx_eth_cap_map); i++) { const struct mcx_eth_proto_capability *cap; if (!ISSET(proto_cap, 1U << i)) continue; cap = &mcx_eth_cap_map[i]; if (cap->cap_media == 0) continue; ifmedia_add(&sc->sc_media, IFM_ETHER | cap->cap_media, 0, NULL); } } static void mcx_media_status(struct ifnet *ifp, struct ifmediareq *ifmr) { struct mcx_softc *sc = (struct mcx_softc *)ifp->if_softc; struct mcx_reg_ptys ptys; int i; uint32_t proto_oper; uint64_t media_oper; memset(&ptys, 0, sizeof(ptys)); ptys.rp_local_port = 1; ptys.rp_proto_mask = MCX_REG_PTYS_PROTO_MASK_ETH; if (mcx_access_hca_reg(sc, MCX_REG_PTYS, MCX_REG_OP_READ, &ptys, sizeof(ptys)) != 0) { printf("%s: unable to read port type/speed\n", DEVNAME(sc)); return; } proto_oper = be32toh(ptys.rp_eth_proto_oper); media_oper = 0; for (i = 0; i < __arraycount(mcx_eth_cap_map); i++) { const struct mcx_eth_proto_capability *cap; if (!ISSET(proto_oper, 1U << i)) continue; cap = &mcx_eth_cap_map[i]; if (cap->cap_media != 0) media_oper = cap->cap_media; } ifmr->ifm_status = IFM_AVALID; if (proto_oper != 0) { ifmr->ifm_status |= IFM_ACTIVE; ifmr->ifm_active = IFM_ETHER | IFM_FDX | IFM_AUTO | media_oper; /* txpause, rxpause, duplex? */ } } static int mcx_media_change(struct ifnet *ifp) { struct mcx_softc *sc = (struct mcx_softc *)ifp->if_softc; struct mcx_reg_ptys ptys; struct mcx_reg_paos paos; uint32_t media; int i, error; if (IFM_TYPE(sc->sc_media.ifm_media) != IFM_ETHER) return EINVAL; error = 0; if (IFM_SUBTYPE(sc->sc_media.ifm_media) == IFM_AUTO) { /* read ptys to get supported media */ memset(&ptys, 0, sizeof(ptys)); ptys.rp_local_port = 1; ptys.rp_proto_mask = MCX_REG_PTYS_PROTO_MASK_ETH; if (mcx_access_hca_reg(sc, MCX_REG_PTYS, MCX_REG_OP_READ, &ptys, sizeof(ptys)) != 0) { printf("%s: unable to read port type/speed\n", DEVNAME(sc)); return EIO; } media = be32toh(ptys.rp_eth_proto_cap); } else { /* map media type */ media = 0; for (i = 0; i < __arraycount(mcx_eth_cap_map); i++) { const struct mcx_eth_proto_capability *cap; cap = &mcx_eth_cap_map[i]; if (cap->cap_media == IFM_SUBTYPE(sc->sc_media.ifm_media)) { media = (1 << i); break; } } } /* disable the port */ memset(&paos, 0, sizeof(paos)); paos.rp_local_port = 1; paos.rp_admin_status = MCX_REG_PAOS_ADMIN_STATUS_DOWN; paos.rp_admin_state_update = MCX_REG_PAOS_ADMIN_STATE_UPDATE_EN; if (mcx_access_hca_reg(sc, MCX_REG_PAOS, MCX_REG_OP_WRITE, &paos, sizeof(paos)) != 0) { printf("%s: unable to set port state to down\n", DEVNAME(sc)); return EIO; } memset(&ptys, 0, sizeof(ptys)); ptys.rp_local_port = 1; ptys.rp_proto_mask = MCX_REG_PTYS_PROTO_MASK_ETH; ptys.rp_eth_proto_admin = htobe32(media); if (mcx_access_hca_reg(sc, MCX_REG_PTYS, MCX_REG_OP_WRITE, &ptys, sizeof(ptys)) != 0) { printf("%s: unable to set port media type/speed\n", DEVNAME(sc)); error = EIO; } /* re-enable the port to start negotiation */ memset(&paos, 0, sizeof(paos)); paos.rp_local_port = 1; paos.rp_admin_status = MCX_REG_PAOS_ADMIN_STATUS_UP; paos.rp_admin_state_update = MCX_REG_PAOS_ADMIN_STATE_UPDATE_EN; if (mcx_access_hca_reg(sc, MCX_REG_PAOS, MCX_REG_OP_WRITE, &paos, sizeof(paos)) != 0) { printf("%s: unable to set port state to up\n", DEVNAME(sc)); error = EIO; } return error; } static void mcx_port_change(struct work *wk, void *xsc) { struct mcx_softc *sc = xsc; struct ifnet *ifp = &sc->sc_ec.ec_if; struct mcx_reg_ptys ptys = { .rp_local_port = 1, .rp_proto_mask = MCX_REG_PTYS_PROTO_MASK_ETH, }; int link_state = LINK_STATE_DOWN; if (mcx_access_hca_reg(sc, MCX_REG_PTYS, MCX_REG_OP_READ, &ptys, sizeof(ptys)) == 0) { uint32_t proto_oper = be32toh(ptys.rp_eth_proto_oper); uint64_t baudrate = 0; unsigned int i; if (proto_oper != 0) link_state = LINK_STATE_UP; for (i = 0; i < __arraycount(mcx_eth_cap_map); i++) { const struct mcx_eth_proto_capability *cap; if (!ISSET(proto_oper, 1U << i)) continue; cap = &mcx_eth_cap_map[i]; if (cap->cap_baudrate == 0) continue; baudrate = cap->cap_baudrate; break; } ifp->if_baudrate = baudrate; } if (link_state != ifp->if_link_state) { if_link_state_change(ifp, link_state); } } static inline uint32_t mcx_rd(struct mcx_softc *sc, bus_size_t r) { uint32_t word; word = bus_space_read_4(sc->sc_memt, sc->sc_memh, r); return (be32toh(word)); } static inline void mcx_wr(struct mcx_softc *sc, bus_size_t r, uint32_t v) { bus_space_write_4(sc->sc_memt, sc->sc_memh, r, htobe32(v)); } static inline void mcx_bar(struct mcx_softc *sc, bus_size_t r, bus_size_t l, int f) { #ifndef __NetBSD__ bus_space_barrier(sc->sc_memt, sc->sc_memh, r, l, f); #endif } static uint64_t mcx_timer(struct mcx_softc *sc) { uint32_t hi, lo, ni; hi = mcx_rd(sc, MCX_INTERNAL_TIMER_H); for (;;) { lo = mcx_rd(sc, MCX_INTERNAL_TIMER_L); mcx_bar(sc, MCX_INTERNAL_TIMER_L, 8, BUS_SPACE_BARRIER_READ); ni = mcx_rd(sc, MCX_INTERNAL_TIMER_H); if (ni == hi) break; hi = ni; } return (((uint64_t)hi << 32) | (uint64_t)lo); } static int mcx_dmamem_alloc(struct mcx_softc *sc, struct mcx_dmamem *mxm, bus_size_t size, u_int align) { mxm->mxm_size = size; if (bus_dmamap_create(sc->sc_dmat, mxm->mxm_size, 1, mxm->mxm_size, 0, BUS_DMA_WAITOK | BUS_DMA_ALLOCNOW, &mxm->mxm_map) != 0) return (1); if (bus_dmamem_alloc(sc->sc_dmat, mxm->mxm_size, align, 0, &mxm->mxm_seg, 1, &mxm->mxm_nsegs, BUS_DMA_WAITOK) != 0) goto destroy; if (bus_dmamem_map(sc->sc_dmat, &mxm->mxm_seg, mxm->mxm_nsegs, mxm->mxm_size, &mxm->mxm_kva, BUS_DMA_WAITOK | BUS_DMA_COHERENT) != 0) goto free; if (bus_dmamap_load(sc->sc_dmat, mxm->mxm_map, mxm->mxm_kva, mxm->mxm_size, NULL, BUS_DMA_WAITOK) != 0) goto unmap; mcx_dmamem_zero(mxm); return (0); unmap: bus_dmamem_unmap(sc->sc_dmat, mxm->mxm_kva, mxm->mxm_size); free: bus_dmamem_free(sc->sc_dmat, &mxm->mxm_seg, 1); destroy: bus_dmamap_destroy(sc->sc_dmat, mxm->mxm_map); return (1); } static void mcx_dmamem_zero(struct mcx_dmamem *mxm) { memset(MCX_DMA_KVA(mxm), 0, MCX_DMA_LEN(mxm)); } static void mcx_dmamem_free(struct mcx_softc *sc, struct mcx_dmamem *mxm) { bus_dmamap_unload(sc->sc_dmat, mxm->mxm_map); bus_dmamem_unmap(sc->sc_dmat, mxm->mxm_kva, mxm->mxm_size); bus_dmamem_free(sc->sc_dmat, &mxm->mxm_seg, 1); bus_dmamap_destroy(sc->sc_dmat, mxm->mxm_map); } static int mcx_hwmem_alloc(struct mcx_softc *sc, struct mcx_hwmem *mhm, unsigned int pages) { bus_dma_segment_t *segs; bus_size_t len = pages * MCX_PAGE_SIZE; size_t seglen; segs = kmem_alloc(sizeof(*segs) * pages, KM_SLEEP); seglen = sizeof(*segs) * pages; if (bus_dmamem_alloc(sc->sc_dmat, len, MCX_PAGE_SIZE, 0, segs, pages, &mhm->mhm_seg_count, BUS_DMA_NOWAIT) != 0) goto free_segs; if (mhm->mhm_seg_count < pages) { size_t nseglen; mhm->mhm_segs = kmem_alloc( sizeof(*mhm->mhm_segs) * mhm->mhm_seg_count, KM_SLEEP); nseglen = sizeof(*mhm->mhm_segs) * mhm->mhm_seg_count; memcpy(mhm->mhm_segs, segs, nseglen); kmem_free(segs, seglen); segs = mhm->mhm_segs; seglen = nseglen; } else mhm->mhm_segs = segs; if (bus_dmamap_create(sc->sc_dmat, len, pages, MCX_PAGE_SIZE, MCX_PAGE_SIZE, BUS_DMA_NOWAIT|BUS_DMA_ALLOCNOW /*|BUS_DMA_64BIT*/, &mhm->mhm_map) != 0) goto free_dmamem; if (bus_dmamap_load_raw(sc->sc_dmat, mhm->mhm_map, mhm->mhm_segs, mhm->mhm_seg_count, len, BUS_DMA_NOWAIT) != 0) goto destroy; bus_dmamap_sync(sc->sc_dmat, mhm->mhm_map, 0, mhm->mhm_map->dm_mapsize, BUS_DMASYNC_PRERW); mhm->mhm_npages = pages; return (0); destroy: bus_dmamap_destroy(sc->sc_dmat, mhm->mhm_map); free_dmamem: bus_dmamem_free(sc->sc_dmat, mhm->mhm_segs, mhm->mhm_seg_count); free_segs: kmem_free(segs, seglen); mhm->mhm_segs = NULL; return (-1); } static void mcx_hwmem_free(struct mcx_softc *sc, struct mcx_hwmem *mhm) { if (mhm->mhm_npages == 0) return; bus_dmamap_sync(sc->sc_dmat, mhm->mhm_map, 0, mhm->mhm_map->dm_mapsize, BUS_DMASYNC_POSTRW); bus_dmamap_unload(sc->sc_dmat, mhm->mhm_map); bus_dmamap_destroy(sc->sc_dmat, mhm->mhm_map); bus_dmamem_free(sc->sc_dmat, mhm->mhm_segs, mhm->mhm_seg_count); kmem_free(mhm->mhm_segs, sizeof(*mhm->mhm_segs) * mhm->mhm_seg_count); mhm->mhm_npages = 0; } #if NKSTAT > 0 struct mcx_ppcnt { char name[KSTAT_KV_NAMELEN]; enum kstat_kv_unit unit; }; static const struct mcx_ppcnt mcx_ppcnt_ieee8023_tpl[] = { { "Good Tx", KSTAT_KV_U_PACKETS, }, { "Good Rx", KSTAT_KV_U_PACKETS, }, { "FCS errs", KSTAT_KV_U_PACKETS, }, { "Alignment Errs", KSTAT_KV_U_PACKETS, }, { "Good Tx", KSTAT_KV_U_BYTES, }, { "Good Rx", KSTAT_KV_U_BYTES, }, { "Multicast Tx", KSTAT_KV_U_PACKETS, }, { "Broadcast Tx", KSTAT_KV_U_PACKETS, }, { "Multicast Rx", KSTAT_KV_U_PACKETS, }, { "Broadcast Rx", KSTAT_KV_U_PACKETS, }, { "In Range Len", KSTAT_KV_U_PACKETS, }, { "Out Of Range Len", KSTAT_KV_U_PACKETS, }, { "Frame Too Long", KSTAT_KV_U_PACKETS, }, { "Symbol Errs", KSTAT_KV_U_PACKETS, }, { "MAC Ctrl Tx", KSTAT_KV_U_PACKETS, }, { "MAC Ctrl Rx", KSTAT_KV_U_PACKETS, }, { "MAC Ctrl Unsup", KSTAT_KV_U_PACKETS, }, { "Pause Rx", KSTAT_KV_U_PACKETS, }, { "Pause Tx", KSTAT_KV_U_PACKETS, }, }; CTASSERT(__arraycount(mcx_ppcnt_ieee8023_tpl) == mcx_ppcnt_ieee8023_count); static const struct mcx_ppcnt mcx_ppcnt_rfc2863_tpl[] = { { "Rx Bytes", KSTAT_KV_U_BYTES, }, { "Rx Unicast", KSTAT_KV_U_PACKETS, }, { "Rx Discards", KSTAT_KV_U_PACKETS, }, { "Rx Errors", KSTAT_KV_U_PACKETS, }, { "Rx Unknown Proto", KSTAT_KV_U_PACKETS, }, { "Tx Bytes", KSTAT_KV_U_BYTES, }, { "Tx Unicast", KSTAT_KV_U_PACKETS, }, { "Tx Discards", KSTAT_KV_U_PACKETS, }, { "Tx Errors", KSTAT_KV_U_PACKETS, }, { "Rx Multicast", KSTAT_KV_U_PACKETS, }, { "Rx Broadcast", KSTAT_KV_U_PACKETS, }, { "Tx Multicast", KSTAT_KV_U_PACKETS, }, { "Tx Broadcast", KSTAT_KV_U_PACKETS, }, }; CTASSERT(__arraycount(mcx_ppcnt_rfc2863_tpl) == mcx_ppcnt_rfc2863_count); static const struct mcx_ppcnt mcx_ppcnt_rfc2819_tpl[] = { { "Drop Events", KSTAT_KV_U_PACKETS, }, { "Octets", KSTAT_KV_U_BYTES, }, { "Packets", KSTAT_KV_U_PACKETS, }, { "Broadcasts", KSTAT_KV_U_PACKETS, }, { "Multicasts", KSTAT_KV_U_PACKETS, }, { "CRC Align Errs", KSTAT_KV_U_PACKETS, }, { "Undersize", KSTAT_KV_U_PACKETS, }, { "Oversize", KSTAT_KV_U_PACKETS, }, { "Fragments", KSTAT_KV_U_PACKETS, }, { "Jabbers", KSTAT_KV_U_PACKETS, }, { "Collisions", KSTAT_KV_U_NONE, }, { "64B", KSTAT_KV_U_PACKETS, }, { "65-127B", KSTAT_KV_U_PACKETS, }, { "128-255B", KSTAT_KV_U_PACKETS, }, { "256-511B", KSTAT_KV_U_PACKETS, }, { "512-1023B", KSTAT_KV_U_PACKETS, }, { "1024-1518B", KSTAT_KV_U_PACKETS, }, { "1519-2047B", KSTAT_KV_U_PACKETS, }, { "2048-4095B", KSTAT_KV_U_PACKETS, }, { "4096-8191B", KSTAT_KV_U_PACKETS, }, { "8192-10239B", KSTAT_KV_U_PACKETS, }, }; CTASSERT(__arraycount(mcx_ppcnt_rfc2819_tpl) == mcx_ppcnt_rfc2819_count); static const struct mcx_ppcnt mcx_ppcnt_rfc3635_tpl[] = { { "Alignment Errs", KSTAT_KV_U_PACKETS, }, { "FCS Errs", KSTAT_KV_U_PACKETS, }, { "Single Colls", KSTAT_KV_U_PACKETS, }, { "Multiple Colls", KSTAT_KV_U_PACKETS, }, { "SQE Test Errs", KSTAT_KV_U_NONE, }, { "Deferred Tx", KSTAT_KV_U_PACKETS, }, { "Late Colls", KSTAT_KV_U_NONE, }, { "Exess Colls", KSTAT_KV_U_NONE, }, { "Int MAC Tx Errs", KSTAT_KV_U_PACKETS, }, { "CSM Sense Errs", KSTAT_KV_U_NONE, }, { "Too Long", KSTAT_KV_U_PACKETS, }, { "Int MAC Rx Errs", KSTAT_KV_U_PACKETS, }, { "Symbol Errs", KSTAT_KV_U_NONE, }, { "Unknown Control", KSTAT_KV_U_PACKETS, }, { "Pause Rx", KSTAT_KV_U_PACKETS, }, { "Pause Tx", KSTAT_KV_U_PACKETS, }, }; CTASSERT(__arraycount(mcx_ppcnt_rfc3635_tpl) == mcx_ppcnt_rfc3635_count); struct mcx_kstat_ppcnt { const char *ksp_name; const struct mcx_ppcnt *ksp_tpl; unsigned int ksp_n; uint8_t ksp_grp; }; static const struct mcx_kstat_ppcnt mcx_kstat_ppcnt_ieee8023 = { .ksp_name = "ieee802.3", .ksp_tpl = mcx_ppcnt_ieee8023_tpl, .ksp_n = __arraycount(mcx_ppcnt_ieee8023_tpl), .ksp_grp = MCX_REG_PPCNT_GRP_IEEE8023, }; static const struct mcx_kstat_ppcnt mcx_kstat_ppcnt_rfc2863 = { .ksp_name = "rfc2863", .ksp_tpl = mcx_ppcnt_rfc2863_tpl, .ksp_n = __arraycount(mcx_ppcnt_rfc2863_tpl), .ksp_grp = MCX_REG_PPCNT_GRP_RFC2863, }; static const struct mcx_kstat_ppcnt mcx_kstat_ppcnt_rfc2819 = { .ksp_name = "rfc2819", .ksp_tpl = mcx_ppcnt_rfc2819_tpl, .ksp_n = __arraycount(mcx_ppcnt_rfc2819_tpl), .ksp_grp = MCX_REG_PPCNT_GRP_RFC2819, }; static const struct mcx_kstat_ppcnt mcx_kstat_ppcnt_rfc3635 = { .ksp_name = "rfc3635", .ksp_tpl = mcx_ppcnt_rfc3635_tpl, .ksp_n = __arraycount(mcx_ppcnt_rfc3635_tpl), .ksp_grp = MCX_REG_PPCNT_GRP_RFC3635, }; static int mcx_kstat_ppcnt_read(struct kstat *); static void mcx_kstat_attach_tmps(struct mcx_softc *sc); static void mcx_kstat_attach_queues(struct mcx_softc *sc); static struct kstat * mcx_kstat_attach_ppcnt(struct mcx_softc *sc, const struct mcx_kstat_ppcnt *ksp) { struct kstat *ks; struct kstat_kv *kvs; unsigned int i; ks = kstat_create(DEVNAME(sc), 0, ksp->ksp_name, 0, KSTAT_T_KV, 0); if (ks == NULL) return (NULL); kvs = mallocarray(ksp->ksp_n, sizeof(*kvs), M_DEVBUF, M_WAITOK); for (i = 0; i < ksp->ksp_n; i++) { const struct mcx_ppcnt *tpl = &ksp->ksp_tpl[i]; kstat_kv_unit_init(&kvs[i], tpl->name, KSTAT_KV_T_COUNTER64, tpl->unit); } ks->ks_softc = sc; ks->ks_ptr = (void *)ksp; ks->ks_data = kvs; ks->ks_datalen = ksp->ksp_n * sizeof(*kvs); ks->ks_read = mcx_kstat_ppcnt_read; kstat_install(ks); return (ks); } static void mcx_kstat_attach(struct mcx_softc *sc) { sc->sc_kstat_ieee8023 = mcx_kstat_attach_ppcnt(sc, &mcx_kstat_ppcnt_ieee8023); sc->sc_kstat_rfc2863 = mcx_kstat_attach_ppcnt(sc, &mcx_kstat_ppcnt_rfc2863); sc->sc_kstat_rfc2819 = mcx_kstat_attach_ppcnt(sc, &mcx_kstat_ppcnt_rfc2819); sc->sc_kstat_rfc3635 = mcx_kstat_attach_ppcnt(sc, &mcx_kstat_ppcnt_rfc3635); mcx_kstat_attach_tmps(sc); mcx_kstat_attach_queues(sc); } static int mcx_kstat_ppcnt_read(struct kstat *ks) { struct mcx_softc *sc = ks->ks_softc; struct mcx_kstat_ppcnt *ksp = ks->ks_ptr; struct mcx_reg_ppcnt ppcnt = { .ppcnt_grp = ksp->ksp_grp, .ppcnt_local_port = 1, }; struct kstat_kv *kvs = ks->ks_data; uint64_t *vs = (uint64_t *)&ppcnt.ppcnt_counter_set; unsigned int i; int rv; KERNEL_LOCK(); /* XXX */ rv = mcx_access_hca_reg(sc, MCX_REG_PPCNT, MCX_REG_OP_READ, &ppcnt, sizeof(ppcnt)); KERNEL_UNLOCK(); if (rv != 0) return (EIO); nanouptime(&ks->ks_updated); for (i = 0; i < ksp->ksp_n; i++) kstat_kv_u64(&kvs[i]) = bemtoh64(&vs[i]); return (0); } struct mcx_kstat_mtmp { struct kstat_kv ktmp_name; struct kstat_kv ktmp_temperature; struct kstat_kv ktmp_threshold_lo; struct kstat_kv ktmp_threshold_hi; }; static const struct mcx_kstat_mtmp mcx_kstat_mtmp_tpl = { KSTAT_KV_INITIALIZER("name", KSTAT_KV_T_ISTR), KSTAT_KV_INITIALIZER("temperature", KSTAT_KV_T_TEMP), KSTAT_KV_INITIALIZER("lo threshold", KSTAT_KV_T_TEMP), KSTAT_KV_INITIALIZER("hi threshold", KSTAT_KV_T_TEMP), }; static const struct timeval mcx_kstat_mtmp_rate = { 1, 0 }; static int mcx_kstat_mtmp_read(struct kstat *); static void mcx_kstat_attach_tmps(struct mcx_softc *sc) { struct kstat *ks; struct mcx_reg_mcam mcam; struct mcx_reg_mtcap mtcap; struct mcx_kstat_mtmp *ktmp; uint64_t map; unsigned int i, n; memset(&mtcap, 0, sizeof(mtcap)); memset(&mcam, 0, sizeof(mcam)); if (sc->sc_mcam_reg == 0) { /* no management capabilities */ return; } if (mcx_access_hca_reg(sc, MCX_REG_MCAM, MCX_REG_OP_READ, &mcam, sizeof(mcam)) != 0) { /* unable to check management capabilities? */ return; } if (MCX_BITFIELD_BIT(mcam.mcam_feature_cap_mask, MCX_MCAM_FEATURE_CAP_SENSOR_MAP) == 0) { /* no sensor map */ return; } if (mcx_access_hca_reg(sc, MCX_REG_MTCAP, MCX_REG_OP_READ, &mtcap, sizeof(mtcap)) != 0) { /* unable to find temperature sensors */ return; } sc->sc_kstat_mtmp_count = mtcap.mtcap_sensor_count; sc->sc_kstat_mtmp = mallocarray(sc->sc_kstat_mtmp_count, sizeof(*sc->sc_kstat_mtmp), M_DEVBUF, M_WAITOK); n = 0; map = bemtoh64(&mtcap.mtcap_sensor_map); for (i = 0; i < sizeof(map) * NBBY; i++) { if (!ISSET(map, (1ULL << i))) continue; ks = kstat_create(DEVNAME(sc), 0, "temperature", i, KSTAT_T_KV, 0); if (ks == NULL) { /* unable to attach temperature sensor %u, i */ continue; } ktmp = malloc(sizeof(*ktmp), M_DEVBUF, M_WAITOK|M_ZERO); *ktmp = mcx_kstat_mtmp_tpl; ks->ks_data = ktmp; ks->ks_datalen = sizeof(*ktmp); TIMEVAL_TO_TIMESPEC(&mcx_kstat_mtmp_rate, &ks->ks_interval); ks->ks_read = mcx_kstat_mtmp_read; ks->ks_softc = sc; kstat_install(ks); sc->sc_kstat_mtmp[n++] = ks; if (n >= sc->sc_kstat_mtmp_count) break; } } static uint64_t mcx_tmp_to_uK(uint16_t *t) { int64_t mt = (int16_t)bemtoh16(t); /* 0.125 C units */ mt *= 1000000 / 8; /* convert to uC */ mt += 273150000; /* convert to uK */ return (mt); } static int mcx_kstat_mtmp_read(struct kstat *ks) { struct mcx_softc *sc = ks->ks_softc; struct mcx_kstat_mtmp *ktmp = ks->ks_data; struct mcx_reg_mtmp mtmp; int rv; struct timeval updated; TIMESPEC_TO_TIMEVAL(&updated, &ks->ks_updated); if (!ratecheck(&updated, &mcx_kstat_mtmp_rate)) return (0); memset(&mtmp, 0, sizeof(mtmp)); htobem16(&mtmp.mtmp_sensor_index, ks->ks_unit); KERNEL_LOCK(); /* XXX */ rv = mcx_access_hca_reg(sc, MCX_REG_MTMP, MCX_REG_OP_READ, &mtmp, sizeof(mtmp)); KERNEL_UNLOCK(); if (rv != 0) return (EIO); memset(kstat_kv_istr(&ktmp->ktmp_name), 0, sizeof(kstat_kv_istr(&ktmp->ktmp_name))); memcpy(kstat_kv_istr(&ktmp->ktmp_name), mtmp.mtmp_sensor_name, sizeof(mtmp.mtmp_sensor_name)); kstat_kv_temp(&ktmp->ktmp_temperature) = mcx_tmp_to_uK(&mtmp.mtmp_temperature); kstat_kv_temp(&ktmp->ktmp_threshold_lo) = mcx_tmp_to_uK(&mtmp.mtmp_temperature_threshold_lo); kstat_kv_temp(&ktmp->ktmp_threshold_hi) = mcx_tmp_to_uK(&mtmp.mtmp_temperature_threshold_hi); TIMEVAL_TO_TIMESPEC(&updated, &ks->ks_updated); return (0); } struct mcx_queuestat { char name[KSTAT_KV_NAMELEN]; enum kstat_kv_type type; }; static const struct mcx_queuestat mcx_queue_kstat_tpl[] = { { "RQ SW prod", KSTAT_KV_T_COUNTER64 }, { "RQ HW prod", KSTAT_KV_T_COUNTER64 }, { "RQ HW cons", KSTAT_KV_T_COUNTER64 }, { "RQ HW state", KSTAT_KV_T_ISTR }, { "SQ SW prod", KSTAT_KV_T_COUNTER64 }, { "SQ SW cons", KSTAT_KV_T_COUNTER64 }, { "SQ HW prod", KSTAT_KV_T_COUNTER64 }, { "SQ HW cons", KSTAT_KV_T_COUNTER64 }, { "SQ HW state", KSTAT_KV_T_ISTR }, { "CQ SW cons", KSTAT_KV_T_COUNTER64 }, { "CQ HW prod", KSTAT_KV_T_COUNTER64 }, { "CQ HW cons", KSTAT_KV_T_COUNTER64 }, { "CQ HW notify", KSTAT_KV_T_COUNTER64 }, { "CQ HW solicit", KSTAT_KV_T_COUNTER64 }, { "CQ HW status", KSTAT_KV_T_ISTR }, { "CQ HW state", KSTAT_KV_T_ISTR }, { "EQ SW cons", KSTAT_KV_T_COUNTER64 }, { "EQ HW prod", KSTAT_KV_T_COUNTER64 }, { "EQ HW cons", KSTAT_KV_T_COUNTER64 }, { "EQ HW status", KSTAT_KV_T_ISTR }, { "EQ HW state", KSTAT_KV_T_ISTR }, }; static int mcx_kstat_queue_read(struct kstat *); static void mcx_kstat_attach_queues(struct mcx_softc *sc) { struct kstat *ks; struct kstat_kv *kvs; int q, i; for (q = 0; q < sc->sc_nqueues; q++) { ks = kstat_create(DEVNAME(sc), 0, "mcx-queues", q, KSTAT_T_KV, 0); if (ks == NULL) { /* unable to attach queue stats %u, q */ continue; } kvs = mallocarray(nitems(mcx_queue_kstat_tpl), sizeof(*kvs), M_DEVBUF, M_WAITOK); for (i = 0; i < nitems(mcx_queue_kstat_tpl); i++) { const struct mcx_queuestat *tpl = &mcx_queue_kstat_tpl[i]; kstat_kv_init(&kvs[i], tpl->name, tpl->type); } ks->ks_softc = &sc->sc_queues[q]; ks->ks_data = kvs; ks->ks_datalen = nitems(mcx_queue_kstat_tpl) * sizeof(*kvs); ks->ks_read = mcx_kstat_queue_read; sc->sc_queues[q].q_kstat = ks; kstat_install(ks); } } static int mcx_kstat_queue_read(struct kstat *ks) { struct mcx_queues *q = ks->ks_softc; struct mcx_softc *sc = q->q_sc; struct kstat_kv *kvs = ks->ks_data; union { struct mcx_rq_ctx rq; struct mcx_sq_ctx sq; struct mcx_cq_ctx cq; struct mcx_eq_ctx eq; } u; const char *text; int error = 0; KERNEL_LOCK(); if (mcx_query_rq(sc, &q->q_rx, &u.rq) != 0) { error = EIO; goto out; } kstat_kv_u64(kvs++) = q->q_rx.rx_prod; kstat_kv_u64(kvs++) = bemtoh32(&u.rq.rq_wq.wq_sw_counter); kstat_kv_u64(kvs++) = bemtoh32(&u.rq.rq_wq.wq_hw_counter); switch ((bemtoh32(&u.rq.rq_flags) & MCX_RQ_CTX_STATE_MASK) >> MCX_RQ_CTX_STATE_SHIFT) { case MCX_RQ_CTX_STATE_RST: text = "RST"; break; case MCX_RQ_CTX_STATE_RDY: text = "RDY"; break; case MCX_RQ_CTX_STATE_ERR: text = "ERR"; break; default: text = "unknown"; break; } strlcpy(kstat_kv_istr(kvs), text, sizeof(kstat_kv_istr(kvs))); kvs++; if (mcx_query_sq(sc, &q->q_tx, &u.sq) != 0) { error = EIO; goto out; } kstat_kv_u64(kvs++) = q->q_tx.tx_prod; kstat_kv_u64(kvs++) = q->q_tx.tx_cons; kstat_kv_u64(kvs++) = bemtoh32(&u.sq.sq_wq.wq_sw_counter); kstat_kv_u64(kvs++) = bemtoh32(&u.sq.sq_wq.wq_hw_counter); switch ((bemtoh32(&u.sq.sq_flags) & MCX_SQ_CTX_STATE_MASK) >> MCX_SQ_CTX_STATE_SHIFT) { case MCX_SQ_CTX_STATE_RST: text = "RST"; break; case MCX_SQ_CTX_STATE_RDY: text = "RDY"; break; case MCX_SQ_CTX_STATE_ERR: text = "ERR"; break; default: text = "unknown"; break; } strlcpy(kstat_kv_istr(kvs), text, sizeof(kstat_kv_istr(kvs))); kvs++; if (mcx_query_cq(sc, &q->q_cq, &u.cq) != 0) { error = EIO; goto out; } kstat_kv_u64(kvs++) = q->q_cq.cq_cons; kstat_kv_u64(kvs++) = bemtoh32(&u.cq.cq_producer_counter); kstat_kv_u64(kvs++) = bemtoh32(&u.cq.cq_consumer_counter); kstat_kv_u64(kvs++) = bemtoh32(&u.cq.cq_last_notified); kstat_kv_u64(kvs++) = bemtoh32(&u.cq.cq_last_solicit); switch ((bemtoh32(&u.cq.cq_status) & MCX_CQ_CTX_STATUS_MASK) >> MCX_CQ_CTX_STATUS_SHIFT) { case MCX_CQ_CTX_STATUS_OK: text = "OK"; break; case MCX_CQ_CTX_STATUS_OVERFLOW: text = "overflow"; break; case MCX_CQ_CTX_STATUS_WRITE_FAIL: text = "write fail"; break; default: text = "unknown"; break; } strlcpy(kstat_kv_istr(kvs), text, sizeof(kstat_kv_istr(kvs))); kvs++; switch ((bemtoh32(&u.cq.cq_status) & MCX_CQ_CTX_STATE_MASK) >> MCX_CQ_CTX_STATE_SHIFT) { case MCX_CQ_CTX_STATE_SOLICITED: text = "solicited"; break; case MCX_CQ_CTX_STATE_ARMED: text = "armed"; break; case MCX_CQ_CTX_STATE_FIRED: text = "fired"; break; default: text = "unknown"; break; } strlcpy(kstat_kv_istr(kvs), text, sizeof(kstat_kv_istr(kvs))); kvs++; if (mcx_query_eq(sc, &q->q_eq, &u.eq) != 0) { error = EIO; goto out; } kstat_kv_u64(kvs++) = q->q_eq.eq_cons; kstat_kv_u64(kvs++) = bemtoh32(&u.eq.eq_producer_counter); kstat_kv_u64(kvs++) = bemtoh32(&u.eq.eq_consumer_counter); switch ((bemtoh32(&u.eq.eq_status) & MCX_EQ_CTX_STATUS_MASK) >> MCX_EQ_CTX_STATUS_SHIFT) { case MCX_EQ_CTX_STATUS_EQ_WRITE_FAILURE: text = "write fail"; break; case MCX_EQ_CTX_STATUS_OK: text = "OK"; break; default: text = "unknown"; break; } strlcpy(kstat_kv_istr(kvs), text, sizeof(kstat_kv_istr(kvs))); kvs++; switch ((bemtoh32(&u.eq.eq_status) & MCX_EQ_CTX_STATE_MASK) >> MCX_EQ_CTX_STATE_SHIFT) { case MCX_EQ_CTX_STATE_ARMED: text = "armed"; break; case MCX_EQ_CTX_STATE_FIRED: text = "fired"; break; default: text = "unknown"; break; } strlcpy(kstat_kv_istr(kvs), text, sizeof(kstat_kv_istr(kvs))); kvs++; nanouptime(&ks->ks_updated); out: KERNEL_UNLOCK(); return (error); } #endif /* NKSTAT > 0 */ static unsigned int mcx_timecounter_read(struct timecounter *tc) { struct mcx_softc *sc = tc->tc_priv; return (mcx_rd(sc, MCX_INTERNAL_TIMER_L)); } static void mcx_timecounter_attach(struct mcx_softc *sc) { struct timecounter *tc = &sc->sc_timecounter; tc->tc_get_timecount = mcx_timecounter_read; tc->tc_counter_mask = ~0U; tc->tc_frequency = sc->sc_khz * 1000; tc->tc_name = device_xname(sc->sc_dev); tc->tc_quality = -100; tc->tc_priv = sc; tc_init(tc); }