Commit 4ee656bb authored by Tony Nguyen's avatar Tony Nguyen Committed by Jeff Kirsher

ice: Trivial fixes

This is a collection of trivial fixes including fixing whitespace, typos,
function headers, reverse Christmas tree, etc.
Signed-off-by: default avatarTony Nguyen <anthony.l.nguyen@intel.com>
Tested-by: default avatarAndrew Bowers <andrewx.bowers@intel.com>
Signed-off-by: default avatarJeff Kirsher <jeffrey.t.kirsher@intel.com>
parent 1d8bd992
...@@ -1660,6 +1660,7 @@ struct ice_aqc_get_pkg_info_resp { ...@@ -1660,6 +1660,7 @@ struct ice_aqc_get_pkg_info_resp {
__le32 count; __le32 count;
struct ice_aqc_get_pkg_info pkg_info[1]; struct ice_aqc_get_pkg_info pkg_info[1];
}; };
/** /**
* struct ice_aq_desc - Admin Queue (AQ) descriptor * struct ice_aq_desc - Admin Queue (AQ) descriptor
* @flags: ICE_AQ_FLAG_* flags * @flags: ICE_AQ_FLAG_* flags
......
...@@ -588,10 +588,10 @@ void ice_output_fw_log(struct ice_hw *hw, struct ice_aq_desc *desc, void *buf) ...@@ -588,10 +588,10 @@ void ice_output_fw_log(struct ice_hw *hw, struct ice_aq_desc *desc, void *buf)
} }
/** /**
* ice_get_itr_intrl_gran - determine int/intrl granularity * ice_get_itr_intrl_gran
* @hw: pointer to the HW struct * @hw: pointer to the HW struct
* *
* Determines the ITR/intrl granularities based on the maximum aggregate * Determines the ITR/INTRL granularities based on the maximum aggregate
* bandwidth according to the device's configuration during power-on. * bandwidth according to the device's configuration during power-on.
*/ */
static void ice_get_itr_intrl_gran(struct ice_hw *hw) static void ice_get_itr_intrl_gran(struct ice_hw *hw)
......
...@@ -1323,13 +1323,13 @@ enum ice_status ice_set_dcb_cfg(struct ice_port_info *pi) ...@@ -1323,13 +1323,13 @@ enum ice_status ice_set_dcb_cfg(struct ice_port_info *pi)
} }
/** /**
* ice_aq_query_port_ets - query port ets configuration * ice_aq_query_port_ets - query port ETS configuration
* @pi: port information structure * @pi: port information structure
* @buf: pointer to buffer * @buf: pointer to buffer
* @buf_size: buffer size in bytes * @buf_size: buffer size in bytes
* @cd: pointer to command details structure or NULL * @cd: pointer to command details structure or NULL
* *
* query current port ets configuration * query current port ETS configuration
*/ */
static enum ice_status static enum ice_status
ice_aq_query_port_ets(struct ice_port_info *pi, ice_aq_query_port_ets(struct ice_port_info *pi,
...@@ -1416,13 +1416,13 @@ ice_update_port_tc_tree_cfg(struct ice_port_info *pi, ...@@ -1416,13 +1416,13 @@ ice_update_port_tc_tree_cfg(struct ice_port_info *pi,
} }
/** /**
* ice_query_port_ets - query port ets configuration * ice_query_port_ets - query port ETS configuration
* @pi: port information structure * @pi: port information structure
* @buf: pointer to buffer * @buf: pointer to buffer
* @buf_size: buffer size in bytes * @buf_size: buffer size in bytes
* @cd: pointer to command details structure or NULL * @cd: pointer to command details structure or NULL
* *
* query current port ets configuration and update the * query current port ETS configuration and update the
* SW DB with the TC changes * SW DB with the TC changes
*/ */
enum ice_status enum ice_status
......
...@@ -412,9 +412,9 @@ static int ice_dcb_init_cfg(struct ice_pf *pf, bool locked) ...@@ -412,9 +412,9 @@ static int ice_dcb_init_cfg(struct ice_pf *pf, bool locked)
} }
/** /**
* ice_dcb_sw_default_config - Apply a default DCB config * ice_dcb_sw_dflt_cfg - Apply a default DCB config
* @pf: PF to apply config to * @pf: PF to apply config to
* @ets_willing: configure ets willing * @ets_willing: configure ETS willing
* @locked: was this function called with RTNL held * @locked: was this function called with RTNL held
*/ */
static int ice_dcb_sw_dflt_cfg(struct ice_pf *pf, bool ets_willing, bool locked) static int ice_dcb_sw_dflt_cfg(struct ice_pf *pf, bool ets_willing, bool locked)
......
...@@ -1230,7 +1230,8 @@ static void ice_vsi_set_rss_flow_fld(struct ice_vsi *vsi) ...@@ -1230,7 +1230,8 @@ static void ice_vsi_set_rss_flow_fld(struct ice_vsi *vsi)
* *
* Returns 0 on success or ENOMEM on failure. * Returns 0 on success or ENOMEM on failure.
*/ */
int ice_add_mac_to_list(struct ice_vsi *vsi, struct list_head *add_list, int
ice_add_mac_to_list(struct ice_vsi *vsi, struct list_head *add_list,
const u8 *macaddr) const u8 *macaddr)
{ {
struct ice_fltr_list_entry *tmp; struct ice_fltr_list_entry *tmp;
...@@ -2824,8 +2825,8 @@ static void ice_vsi_update_q_map(struct ice_vsi *vsi, struct ice_vsi_ctx *ctx) ...@@ -2824,8 +2825,8 @@ static void ice_vsi_update_q_map(struct ice_vsi *vsi, struct ice_vsi_ctx *ctx)
int ice_vsi_cfg_tc(struct ice_vsi *vsi, u8 ena_tc) int ice_vsi_cfg_tc(struct ice_vsi *vsi, u8 ena_tc)
{ {
u16 max_txqs[ICE_MAX_TRAFFIC_CLASS] = { 0 }; u16 max_txqs[ICE_MAX_TRAFFIC_CLASS] = { 0 };
struct ice_vsi_ctx *ctx;
struct ice_pf *pf = vsi->back; struct ice_pf *pf = vsi->back;
struct ice_vsi_ctx *ctx;
enum ice_status status; enum ice_status status;
struct device *dev; struct device *dev;
int i, ret = 0; int i, ret = 0;
......
...@@ -3153,7 +3153,9 @@ ice_probe(struct pci_dev *pdev, const struct pci_device_id __always_unused *ent) ...@@ -3153,7 +3153,9 @@ ice_probe(struct pci_dev *pdev, const struct pci_device_id __always_unused *ent)
struct ice_hw *hw; struct ice_hw *hw;
int err; int err;
/* this driver uses devres, see Documentation/driver-api/driver-model/devres.rst */ /* this driver uses devres, see
* Documentation/driver-api/driver-model/devres.rst
*/
err = pcim_enable_device(pdev); err = pcim_enable_device(pdev);
if (err) if (err)
return err; return err;
......
...@@ -644,7 +644,7 @@ static bool ice_page_is_reserved(struct page *page) ...@@ -644,7 +644,7 @@ static bool ice_page_is_reserved(struct page *page)
* Update the offset within page so that Rx buf will be ready to be reused. * Update the offset within page so that Rx buf will be ready to be reused.
* For systems with PAGE_SIZE < 8192 this function will flip the page offset * For systems with PAGE_SIZE < 8192 this function will flip the page offset
* so the second half of page assigned to Rx buffer will be used, otherwise * so the second half of page assigned to Rx buffer will be used, otherwise
* the offset is moved by the @size bytes * the offset is moved by "size" bytes
*/ */
static void static void
ice_rx_buf_adjust_pg_offset(struct ice_rx_buf *rx_buf, unsigned int size) ice_rx_buf_adjust_pg_offset(struct ice_rx_buf *rx_buf, unsigned int size)
...@@ -1619,11 +1619,11 @@ ice_tx_map(struct ice_ring *tx_ring, struct ice_tx_buf *first, ...@@ -1619,11 +1619,11 @@ ice_tx_map(struct ice_ring *tx_ring, struct ice_tx_buf *first,
{ {
u64 td_offset, td_tag, td_cmd; u64 td_offset, td_tag, td_cmd;
u16 i = tx_ring->next_to_use; u16 i = tx_ring->next_to_use;
skb_frag_t *frag;
unsigned int data_len, size; unsigned int data_len, size;
struct ice_tx_desc *tx_desc; struct ice_tx_desc *tx_desc;
struct ice_tx_buf *tx_buf; struct ice_tx_buf *tx_buf;
struct sk_buff *skb; struct sk_buff *skb;
skb_frag_t *frag;
dma_addr_t dma; dma_addr_t dma;
td_tag = off->td_l2tag1; td_tag = off->td_l2tag1;
...@@ -1736,9 +1736,8 @@ ice_tx_map(struct ice_ring *tx_ring, struct ice_tx_buf *first, ...@@ -1736,9 +1736,8 @@ ice_tx_map(struct ice_ring *tx_ring, struct ice_tx_buf *first,
ice_maybe_stop_tx(tx_ring, DESC_NEEDED); ice_maybe_stop_tx(tx_ring, DESC_NEEDED);
/* notify HW of packet */ /* notify HW of packet */
if (netif_xmit_stopped(txring_txq(tx_ring)) || !netdev_xmit_more()) { if (netif_xmit_stopped(txring_txq(tx_ring)) || !netdev_xmit_more())
writel(i, tx_ring->tail); writel(i, tx_ring->tail);
}
return; return;
...@@ -2076,7 +2075,7 @@ static bool __ice_chk_linearize(struct sk_buff *skb) ...@@ -2076,7 +2075,7 @@ static bool __ice_chk_linearize(struct sk_buff *skb)
frag = &skb_shinfo(skb)->frags[0]; frag = &skb_shinfo(skb)->frags[0];
/* Initialize size to the negative value of gso_size minus 1. We /* Initialize size to the negative value of gso_size minus 1. We
* use this as the worst case scenerio in which the frag ahead * use this as the worst case scenario in which the frag ahead
* of us only provides one byte which is why we are limited to 6 * of us only provides one byte which is why we are limited to 6
* descriptors for a single transmit as the header and previous * descriptors for a single transmit as the header and previous
* fragment are already consuming 2 descriptors. * fragment are already consuming 2 descriptors.
......
...@@ -517,7 +517,7 @@ struct ice_hw { ...@@ -517,7 +517,7 @@ struct ice_hw {
struct ice_fw_log_cfg fw_log; struct ice_fw_log_cfg fw_log;
/* Device max aggregate bandwidths corresponding to the GL_PWR_MODE_CTL /* Device max aggregate bandwidths corresponding to the GL_PWR_MODE_CTL
* register. Used for determining the ITR/intrl granularity during * register. Used for determining the ITR/INTRL granularity during
* initialization. * initialization.
*/ */
#define ICE_MAX_AGG_BW_200G 0x0 #define ICE_MAX_AGG_BW_200G 0x0
......
...@@ -1093,7 +1093,6 @@ bool ice_reset_all_vfs(struct ice_pf *pf, bool is_vflr) ...@@ -1093,7 +1093,6 @@ bool ice_reset_all_vfs(struct ice_pf *pf, bool is_vflr)
* finished resetting. * finished resetting.
*/ */
for (i = 0, v = 0; i < 10 && v < pf->num_alloc_vfs; i++) { for (i = 0, v = 0; i < 10 && v < pf->num_alloc_vfs; i++) {
/* Check each VF in sequence */ /* Check each VF in sequence */
while (v < pf->num_alloc_vfs) { while (v < pf->num_alloc_vfs) {
u32 reg; u32 reg;
...@@ -2637,8 +2636,8 @@ static int ice_vc_request_qs_msg(struct ice_vf *vf, u8 *msg) ...@@ -2637,8 +2636,8 @@ static int ice_vc_request_qs_msg(struct ice_vf *vf, u8 *msg)
struct ice_pf *pf = vf->pf; struct ice_pf *pf = vf->pf;
u16 max_allowed_vf_queues; u16 max_allowed_vf_queues;
u16 tx_rx_queue_left; u16 tx_rx_queue_left;
u16 cur_queues;
struct device *dev; struct device *dev;
u16 cur_queues;
dev = ice_pf_to_dev(pf); dev = ice_pf_to_dev(pf);
if (!test_bit(ICE_VF_STATE_ACTIVE, vf->vf_states)) { if (!test_bit(ICE_VF_STATE_ACTIVE, vf->vf_states)) {
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment