/* * CDDL HEADER START * * The contents of this file are subject to the terms of the * Common Development and Distribution License, v.1, (the "License"). * You may not use this file except in compliance with the License. * * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE * or http://opensource.org/licenses/CDDL-1.0. * See the License for the specific language governing permissions * and limitations under the License. * * When distributing Covered Code, include this CDDL HEADER in each * file and include the License file at usr/src/OPENSOLARIS.LICENSE. * If applicable, add the following below this CDDL HEADER, with the * fields enclosed by brackets "[]" replaced with your own identifying * information: Portions Copyright [yyyy] [name of copyright owner] * * CDDL HEADER END */ /* * Copyright 2014-2017 Cavium, Inc. * The contents of this file are subject to the terms of the Common Development * and Distribution License, v.1, (the "License"). * You may not use this file except in compliance with the License. * You can obtain a copy of the License at available * at http://opensource.org/licenses/CDDL-1.0 * See the License for the specific language governing permissions and * limitations under the License. */ #ifndef __ECORE_DEV_API_H__ #define __ECORE_DEV_API_H__ #include "ecore_status.h" #include "ecore_chain.h" #include "ecore_int_api.h" struct ecore_wake_info { u32 wk_info; u32 wk_details; u32 wk_pkt_len; u8 wk_buffer[256]; }; /** * @brief ecore_init_dp - initialize the debug level * * @param p_dev * @param dp_module * @param dp_level * @param dp_ctx */ void ecore_init_dp(struct ecore_dev *p_dev, u32 dp_module, u8 dp_level, void *dp_ctx); /** * @brief ecore_init_struct - initialize the device structure to * its defaults * * @param p_dev */ void ecore_init_struct(struct ecore_dev *p_dev); /** * @brief ecore_resc_free - * * @param p_dev */ void ecore_resc_free(struct ecore_dev *p_dev); /** * @brief ecore_resc_alloc - * * @param p_dev * * @return enum _ecore_status_t */ enum _ecore_status_t ecore_resc_alloc(struct ecore_dev *p_dev); /** * @brief ecore_resc_setup - * * @param p_dev */ void ecore_resc_setup(struct ecore_dev *p_dev); enum ecore_override_force_load { ECORE_OVERRIDE_FORCE_LOAD_NONE, ECORE_OVERRIDE_FORCE_LOAD_ALWAYS, ECORE_OVERRIDE_FORCE_LOAD_NEVER, }; struct ecore_drv_load_params { /* Indicates whether the driver is running over a crash kernel. * As part of the load request, this will be used for providing the * driver role to the MFW. * In case of a crash kernel over PDA - this should be set to false. */ bool is_crash_kernel; /* The timeout value that the MFW should use when locking the engine for * the driver load process. * A value of '0' means the default value, and '255' means no timeout. */ u8 mfw_timeout_val; #define ECORE_LOAD_REQ_LOCK_TO_DEFAULT 0 #define ECORE_LOAD_REQ_LOCK_TO_NONE 255 /* Avoid engine reset when first PF loads on it */ bool avoid_eng_reset; /* Allow overriding the default force load behavior */ enum ecore_override_force_load override_force_load; }; struct ecore_hw_init_params { /* Tunneling parameters */ struct ecore_tunnel_info *p_tunn; bool b_hw_start; /* Interrupt mode [msix, inta, etc.] to use */ enum ecore_int_mode int_mode; /* NPAR tx switching to be used for vports configured for tx-switching */ bool allow_npar_tx_switch; /* Binary fw data pointer in binary fw file */ const u8 *bin_fw_data; /* Driver load parameters */ struct ecore_drv_load_params *p_drv_load_params; }; /** * @brief ecore_hw_init - * * @param p_dev * @param p_params * * @return enum _ecore_status_t */ enum _ecore_status_t ecore_hw_init(struct ecore_dev *p_dev, struct ecore_hw_init_params *p_params); /** * @brief ecore_hw_timers_stop_all - * * @param p_dev * * @return void */ void ecore_hw_timers_stop_all(struct ecore_dev *p_dev); /** * @brief ecore_hw_stop - * * @param p_dev * * @return enum _ecore_status_t */ enum _ecore_status_t ecore_hw_stop(struct ecore_dev *p_dev); /** * @brief ecore_hw_stop_fastpath -should be called incase * slowpath is still required for the device, * but fastpath is not. * * @param p_dev * */ void ecore_hw_stop_fastpath(struct ecore_dev *p_dev); #ifndef LINUX_REMOVE /** * @brief ecore_hw_hibernate_prepare -should be called when * the system is going into the hibernate state * * @param p_dev * */ void ecore_hw_hibernate_prepare(struct ecore_dev *p_dev); /** * @brief ecore_hw_hibernate_resume -should be called when the system is resuming from D3 power state and before calling ecore_hw_init. * * @param p_hwfn * */ void ecore_hw_hibernate_resume(struct ecore_dev *p_dev); #endif /** * @brief ecore_hw_start_fastpath -restart fastpath traffic, * only if hw_stop_fastpath was called * @param p_dev * */ void ecore_hw_start_fastpath(struct ecore_hwfn *p_hwfn); enum ecore_hw_prepare_result { ECORE_HW_PREPARE_SUCCESS, /* FAILED results indicate probe has failed & cleaned up */ ECORE_HW_PREPARE_FAILED_ENG2, ECORE_HW_PREPARE_FAILED_ME, ECORE_HW_PREPARE_FAILED_MEM, ECORE_HW_PREPARE_FAILED_DEV, ECORE_HW_PREPARE_FAILED_NVM, /* BAD results indicate probe is passed even though some wrongness * has occurred; Trying to actually use [I.e., hw_init()] might have * dire reprecautions. */ ECORE_HW_PREPARE_BAD_IOV, ECORE_HW_PREPARE_BAD_MCP, ECORE_HW_PREPARE_BAD_IGU, }; struct ecore_hw_prepare_params { /* Personality to initialize */ int personality; /* Force the driver's default resource allocation */ bool drv_resc_alloc; /* Check the reg_fifo after any register access */ bool chk_reg_fifo; /* Request the MFW to initiate PF FLR */ bool initiate_pf_flr; /* The OS Epoch time in seconds */ u32 epoch; /* Allow the MFW to collect a crash dump */ bool allow_mdump; /* Allow prepare to pass even if some initializations are failing. * If set, the `p_prepare_res' field would be set with the return, * and might allow probe to pass even if there are certain issues. */ bool b_relaxed_probe; enum ecore_hw_prepare_result p_relaxed_res; }; /** * @brief ecore_hw_prepare - * * @param p_dev * @param p_params * * @return enum _ecore_status_t */ enum _ecore_status_t ecore_hw_prepare(struct ecore_dev *p_dev, struct ecore_hw_prepare_params *p_params); /** * @brief ecore_hw_remove - * * @param p_dev */ void ecore_hw_remove(struct ecore_dev *p_dev); /** * @brief ecore_set_nwuf_reg - * * @param p_dev * @param wol_flag - wol_capability * @param reg_idx - Index of the pattern register * @param pattern_size - size of pattern * @param crc - CRC value of patter & mask * * @return enum _ecore_status_t */ enum _ecore_status_t ecore_set_nwuf_reg(struct ecore_dev *p_dev, const bool b_enable, u32 reg_idx, u32 pattern_size, u32 crc); /** * @brief ecore_get_wake_info - get magic packet buffer * * @param p_dev * @param wake_info - pointer to ecore_wake_info buffer * * @return enum _ecore_status_t */ enum _ecore_status_t ecore_get_wake_info(struct ecore_dev *p_dev, struct ecore_wake_info *wake_info); /** * @brief ecore_wol_buffer_clear - Clear magic package buffer * * @param p_dev * * @return void */ void ecore_wol_buffer_clear(struct ecore_dev *p_dev); /** * @brief ecore_ptt_acquire - Allocate a PTT window * * Should be called at the entry point to the driver (at the beginning of an * exported function) * * @param p_hwfn * * @return struct ecore_ptt */ struct ecore_ptt *ecore_ptt_acquire(struct ecore_hwfn *p_hwfn); /** * @brief ecore_ptt_release - Release PTT Window * * Should be called at the end of a flow - at the end of the function that * acquired the PTT. * * * @param p_hwfn * @param p_ptt */ void ecore_ptt_release(struct ecore_hwfn *p_hwfn, struct ecore_ptt *p_ptt); #ifndef __EXTRACT__LINUX__ struct ecore_eth_stats_common { u64 no_buff_discards; u64 packet_too_big_discard; u64 ttl0_discard; u64 rx_ucast_bytes; u64 rx_mcast_bytes; u64 rx_bcast_bytes; u64 rx_ucast_pkts; u64 rx_mcast_pkts; u64 rx_bcast_pkts; u64 mftag_filter_discards; u64 mac_filter_discards; u64 tx_ucast_bytes; u64 tx_mcast_bytes; u64 tx_bcast_bytes; u64 tx_ucast_pkts; u64 tx_mcast_pkts; u64 tx_bcast_pkts; u64 tx_err_drop_pkts; u64 tpa_coalesced_pkts; u64 tpa_coalesced_events; u64 tpa_aborts_num; u64 tpa_not_coalesced_pkts; u64 tpa_coalesced_bytes; /* port */ u64 rx_64_byte_packets; u64 rx_65_to_127_byte_packets; u64 rx_128_to_255_byte_packets; u64 rx_256_to_511_byte_packets; u64 rx_512_to_1023_byte_packets; u64 rx_1024_to_1518_byte_packets; u64 rx_crc_errors; u64 rx_mac_crtl_frames; u64 rx_pause_frames; u64 rx_pfc_frames; u64 rx_align_errors; u64 rx_carrier_errors; u64 rx_oversize_packets; u64 rx_jabbers; u64 rx_undersize_packets; u64 rx_fragments; u64 tx_64_byte_packets; u64 tx_65_to_127_byte_packets; u64 tx_128_to_255_byte_packets; u64 tx_256_to_511_byte_packets; u64 tx_512_to_1023_byte_packets; u64 tx_1024_to_1518_byte_packets; u64 tx_pause_frames; u64 tx_pfc_frames; u64 brb_truncates; u64 brb_discards; u64 rx_mac_bytes; u64 rx_mac_uc_packets; u64 rx_mac_mc_packets; u64 rx_mac_bc_packets; u64 rx_mac_frames_ok; u64 tx_mac_bytes; u64 tx_mac_uc_packets; u64 tx_mac_mc_packets; u64 tx_mac_bc_packets; u64 tx_mac_ctrl_frames; }; struct ecore_eth_stats_bb { u64 rx_1519_to_1522_byte_packets; u64 rx_1519_to_2047_byte_packets; u64 rx_2048_to_4095_byte_packets; u64 rx_4096_to_9216_byte_packets; u64 rx_9217_to_16383_byte_packets; u64 tx_1519_to_2047_byte_packets; u64 tx_2048_to_4095_byte_packets; u64 tx_4096_to_9216_byte_packets; u64 tx_9217_to_16383_byte_packets; u64 tx_lpi_entry_count; u64 tx_total_collisions; }; struct ecore_eth_stats_ah { u64 rx_1519_to_max_byte_packets; u64 tx_1519_to_max_byte_packets; }; struct ecore_eth_stats { struct ecore_eth_stats_common common; union { struct ecore_eth_stats_bb bb; struct ecore_eth_stats_ah ah; }; }; #endif enum ecore_dmae_address_type_t { ECORE_DMAE_ADDRESS_HOST_VIRT, ECORE_DMAE_ADDRESS_HOST_PHYS, ECORE_DMAE_ADDRESS_GRC }; /* value of flags If ECORE_DMAE_FLAG_RW_REPL_SRC flag is set and the * source is a block of length DMAE_MAX_RW_SIZE and the * destination is larger, the source block will be duplicated as * many times as required to fill the destination block. This is * used mostly to write a zeroed buffer to destination address * using DMA */ #define ECORE_DMAE_FLAG_RW_REPL_SRC 0x00000001 #define ECORE_DMAE_FLAG_VF_SRC 0x00000002 #define ECORE_DMAE_FLAG_VF_DST 0x00000004 #define ECORE_DMAE_FLAG_COMPLETION_DST 0x00000008 struct ecore_dmae_params { u32 flags; /* consists of ECORE_DMAE_FLAG_* values */ u8 src_vfid; u8 dst_vfid; }; /** * @brief ecore_dmae_host2grc - copy data from source addr to * dmae registers using the given ptt * * @param p_hwfn * @param p_ptt * @param source_addr * @param grc_addr (dmae_data_offset) * @param size_in_dwords * @param flags (one of the flags defined above) */ enum _ecore_status_t ecore_dmae_host2grc(struct ecore_hwfn *p_hwfn, struct ecore_ptt *p_ptt, u64 source_addr, u32 grc_addr, u32 size_in_dwords, u32 flags); /** * @brief ecore_dmae_grc2host - Read data from dmae data offset * to source address using the given ptt * * @param p_ptt * @param grc_addr (dmae_data_offset) * @param dest_addr * @param size_in_dwords * @param flags - one of the flags defined above */ enum _ecore_status_t ecore_dmae_grc2host(struct ecore_hwfn *p_hwfn, struct ecore_ptt *p_ptt, u32 grc_addr, dma_addr_t dest_addr, u32 size_in_dwords, u32 flags); /** * @brief ecore_dmae_host2host - copy data from to source address * to a destination adress (for SRIOV) using the given ptt * * @param p_hwfn * @param p_ptt * @param source_addr * @param dest_addr * @param size_in_dwords * @param params */ enum _ecore_status_t ecore_dmae_host2host(struct ecore_hwfn *p_hwfn, struct ecore_ptt *p_ptt, dma_addr_t source_addr, dma_addr_t dest_addr, u32 size_in_dwords, struct ecore_dmae_params *p_params); /** * @brief ecore_chain_alloc - Allocate and initialize a chain * * @param p_hwfn * @param intended_use * @param mode * @param num_elems * @param elem_size * @param p_chain * * @return enum _ecore_status_t */ enum _ecore_status_t ecore_chain_alloc(struct ecore_dev *p_dev, enum ecore_chain_use_mode intended_use, enum ecore_chain_mode mode, enum ecore_chain_cnt_type cnt_type, u32 num_elems, osal_size_t elem_size, struct ecore_chain *p_chain, struct ecore_chain_ext_pbl *ext_pbl); /** * @brief ecore_chain_free - Free chain DMA memory * * @param p_hwfn * @param p_chain */ void ecore_chain_free(struct ecore_dev *p_dev, struct ecore_chain *p_chain); /** * @@brief ecore_fw_l2_queue - Get absolute L2 queue ID * * @param p_hwfn * @param src_id - relative to p_hwfn * @param dst_id - absolute per engine * * @return enum _ecore_status_t */ enum _ecore_status_t ecore_fw_l2_queue(struct ecore_hwfn *p_hwfn, u16 src_id, u16 *dst_id); /** * @@brief ecore_fw_vport - Get absolute vport ID * * @param p_hwfn * @param src_id - relative to p_hwfn * @param dst_id - absolute per engine * * @return enum _ecore_status_t */ enum _ecore_status_t ecore_fw_vport(struct ecore_hwfn *p_hwfn, u8 src_id, u8 *dst_id); /** * @@brief ecore_fw_rss_eng - Get absolute RSS engine ID * * @param p_hwfn * @param src_id - relative to p_hwfn * @param dst_id - absolute per engine * * @return enum _ecore_status_t */ enum _ecore_status_t ecore_fw_rss_eng(struct ecore_hwfn *p_hwfn, u8 src_id, u8 *dst_id); /** * @brief ecore_llh_add_mac_filter - configures a MAC filter in llh * * @param p_hwfn * @param p_ptt * @param p_filter - MAC to add */ enum _ecore_status_t ecore_llh_add_mac_filter(struct ecore_hwfn *p_hwfn, struct ecore_ptt *p_ptt, u8 *p_filter); /** * @brief ecore_llh_remove_mac_filter - removes a MAC filtre from llh * * @param p_hwfn * @param p_ptt * @param p_filter - MAC to remove */ void ecore_llh_remove_mac_filter(struct ecore_hwfn *p_hwfn, struct ecore_ptt *p_ptt, u8 *p_filter); enum ecore_llh_port_filter_type_t { ECORE_LLH_FILTER_ETHERTYPE, ECORE_LLH_FILTER_TCP_SRC_PORT, ECORE_LLH_FILTER_TCP_DEST_PORT, ECORE_LLH_FILTER_TCP_SRC_AND_DEST_PORT, ECORE_LLH_FILTER_UDP_SRC_PORT, ECORE_LLH_FILTER_UDP_DEST_PORT, ECORE_LLH_FILTER_UDP_SRC_AND_DEST_PORT }; /** * @brief ecore_llh_add_protocol_filter - configures a protocol filter in llh * * @param p_hwfn * @param p_ptt * @param source_port_or_eth_type - source port or ethertype to add * @param dest_port - destination port to add * @param type - type of filters and comparing */ enum _ecore_status_t ecore_llh_add_protocol_filter(struct ecore_hwfn *p_hwfn, struct ecore_ptt *p_ptt, u16 source_port_or_eth_type, u16 dest_port, enum ecore_llh_port_filter_type_t type); /** * @brief ecore_llh_remove_protocol_filter - remove a protocol filter in llh * * @param p_hwfn * @param p_ptt * @param source_port_or_eth_type - source port or ethertype to add * @param dest_port - destination port to add * @param type - type of filters and comparing */ void ecore_llh_remove_protocol_filter(struct ecore_hwfn *p_hwfn, struct ecore_ptt *p_ptt, u16 source_port_or_eth_type, u16 dest_port, enum ecore_llh_port_filter_type_t type); /** * @brief ecore_llh_clear_all_filters - removes all MAC filters from llh * * @param p_hwfn * @param p_ptt */ void ecore_llh_clear_all_filters(struct ecore_hwfn *p_hwfn, struct ecore_ptt *p_ptt); /** * @brief ecore_llh_set_function_as_default - set function as defult per port * * @param p_hwfn * @param p_ptt */ enum _ecore_status_t ecore_llh_set_function_as_default(struct ecore_hwfn *p_hwfn, struct ecore_ptt *p_ptt); /** *@brief Cleanup of previous driver remains prior to load * * @param p_hwfn * @param p_ptt * @param id - For PF, engine-relative. For VF, PF-relative. * @param is_vf - true iff cleanup is made for a VF. * * @return enum _ecore_status_t */ enum _ecore_status_t ecore_final_cleanup(struct ecore_hwfn *p_hwfn, struct ecore_ptt *p_ptt, u16 id, bool is_vf); /** * @brief ecore_set_queue_coalesce - Configure coalesce parameters for Rx and * Tx queue. The fact that we can configure coalescing to up to 511, but on * varying accuracy [the bigger the value the less accurate] up to a mistake * of 3usec for the highest values. * While the API allows setting coalescing per-qid, all queues sharing a SB * should be in same range [i.e., either 0-0x7f, 0x80-0xff or 0x100-0x1ff] * otherwise configuration would break. * * @param p_hwfn * @param rx_coal - Rx Coalesce value in micro seconds. * @param tx_coal - TX Coalesce value in micro seconds. * @param p_handle * * @return enum _ecore_status_t **/ enum _ecore_status_t ecore_set_queue_coalesce(struct ecore_hwfn *p_hwfn, u16 rx_coal, u16 tx_coal, void *p_handle); /** * @brief - Recalculate feature distributions based on HW resources and * user inputs. Currently this affects RDMA_CNQ, PF_L2_QUE and VF_L2_QUE. * As a result, this must not be called while RDMA is active or while VFs * are enabled. * * @param p_hwfn */ void ecore_hw_set_feat(struct ecore_hwfn *p_hwfn); /** * @brief ecore_change_pci_hwfn - Enable or disable PCI BUS MASTER * * @param p_hwfn * @param p_ptt * @param enable - true/false * * @return enum _ecore_status_t */ enum _ecore_status_t ecore_change_pci_hwfn(struct ecore_hwfn *p_hwfn, struct ecore_ptt *p_ptt, u8 enable); #endif