Commit 7eadc0a0 authored by Jakub Kicinski's avatar Jakub Kicinski

Merge tag 'mlx5-updates-2023-02-07' of git://git.kernel.org/pub/scm/linux/kernel/git/saeed/linux

Saeed Mahameed says:

====================
mlx5-updates-2023-02-07

1) Minor and trivial code Cleanups

2) Minor fixes for net-next

3) From Shay: dynamic FW trace strings update.

* tag 'mlx5-updates-2023-02-07' of git://git.kernel.org/pub/scm/linux/kernel/git/saeed/linux:
  net/mlx5: fw_tracer, Add support for unrecognized string
  net/mlx5: fw_tracer, Add support for strings DB update event
  net/mlx5: fw_tracer, allow 0 size string DBs
  net/mlx5: fw_tracer: Fix debug print
  net/mlx5: fs, Remove redundant assignment of size
  net/mlx5: fs_core, Remove redundant variable err
  net/mlx5: Fix memory leak in error flow of port set buffer
  net/mlx5e: Remove incorrect debugfs_create_dir NULL check in TLS
  net/mlx5e: Remove incorrect debugfs_create_dir NULL check in hairpin
  net/mlx5: fs, Remove redundant vport_number assignment
  net/mlx5e: Remove redundant code for handling vlan actions
  net/mlx5e: Don't listen to remove flows event
  net/mlx5: fw reset: Skip device ID check if PCI link up failed
  net/mlx5: Remove redundant health work lock
  mlx5: reduce stack usage in mlx5_setup_tc
====================

Link: https://lore.kernel.org/r/20230208003712.68386-1-saeed@kernel.orgSigned-off-by: default avatarJakub Kicinski <kuba@kernel.org>
parents e6ebe6c1 f7133135
...@@ -233,6 +233,8 @@ static int mlx5_fw_tracer_allocate_strings_db(struct mlx5_fw_tracer *tracer) ...@@ -233,6 +233,8 @@ static int mlx5_fw_tracer_allocate_strings_db(struct mlx5_fw_tracer *tracer)
int i; int i;
for (i = 0; i < num_string_db; i++) { for (i = 0; i < num_string_db; i++) {
if (!string_db_size_out[i])
continue;
tracer->str_db.buffer[i] = kzalloc(string_db_size_out[i], GFP_KERNEL); tracer->str_db.buffer[i] = kzalloc(string_db_size_out[i], GFP_KERNEL);
if (!tracer->str_db.buffer[i]) if (!tracer->str_db.buffer[i])
goto free_strings_db; goto free_strings_db;
...@@ -278,6 +280,8 @@ static void mlx5_tracer_read_strings_db(struct work_struct *work) ...@@ -278,6 +280,8 @@ static void mlx5_tracer_read_strings_db(struct work_struct *work)
} }
for (i = 0; i < num_string_db; i++) { for (i = 0; i < num_string_db; i++) {
if (!tracer->str_db.size_out[i])
continue;
offset = 0; offset = 0;
MLX5_SET(mtrc_stdb, in, string_db_index, i); MLX5_SET(mtrc_stdb, in, string_db_index, i);
num_of_reads = tracer->str_db.size_out[i] / num_of_reads = tracer->str_db.size_out[i] /
...@@ -384,6 +388,8 @@ static struct tracer_string_format *mlx5_tracer_get_string(struct mlx5_fw_tracer ...@@ -384,6 +388,8 @@ static struct tracer_string_format *mlx5_tracer_get_string(struct mlx5_fw_tracer
str_ptr = tracer_event->string_event.string_param; str_ptr = tracer_event->string_event.string_param;
for (i = 0; i < tracer->str_db.num_string_db; i++) { for (i = 0; i < tracer->str_db.num_string_db; i++) {
if (!tracer->str_db.size_out[i])
continue;
if (str_ptr > tracer->str_db.base_address_out[i] && if (str_ptr > tracer->str_db.base_address_out[i] &&
str_ptr < tracer->str_db.base_address_out[i] + str_ptr < tracer->str_db.base_address_out[i] +
tracer->str_db.size_out[i]) { tracer->str_db.size_out[i]) {
...@@ -459,6 +465,7 @@ static void poll_trace(struct mlx5_fw_tracer *tracer, ...@@ -459,6 +465,7 @@ static void poll_trace(struct mlx5_fw_tracer *tracer,
tracer_event->event_id = MLX5_GET(tracer_event, trace, event_id); tracer_event->event_id = MLX5_GET(tracer_event, trace, event_id);
tracer_event->lost_event = MLX5_GET(tracer_event, trace, lost); tracer_event->lost_event = MLX5_GET(tracer_event, trace, lost);
tracer_event->out = trace;
switch (tracer_event->event_id) { switch (tracer_event->event_id) {
case TRACER_EVENT_TYPE_TIMESTAMP: case TRACER_EVENT_TYPE_TIMESTAMP:
...@@ -581,6 +588,26 @@ void mlx5_tracer_print_trace(struct tracer_string_format *str_frmt, ...@@ -581,6 +588,26 @@ void mlx5_tracer_print_trace(struct tracer_string_format *str_frmt,
mlx5_tracer_clean_message(str_frmt); mlx5_tracer_clean_message(str_frmt);
} }
static int mlx5_tracer_handle_raw_string(struct mlx5_fw_tracer *tracer,
struct tracer_event *tracer_event)
{
struct tracer_string_format *cur_string;
cur_string = mlx5_tracer_message_insert(tracer, tracer_event);
if (!cur_string)
return -1;
cur_string->event_id = tracer_event->event_id;
cur_string->timestamp = tracer_event->string_event.timestamp;
cur_string->lost = tracer_event->lost_event;
cur_string->string = "0x%08x%08x";
cur_string->num_of_params = 2;
cur_string->params[0] = upper_32_bits(*tracer_event->out);
cur_string->params[1] = lower_32_bits(*tracer_event->out);
list_add_tail(&cur_string->list, &tracer->ready_strings_list);
return 0;
}
static int mlx5_tracer_handle_string_trace(struct mlx5_fw_tracer *tracer, static int mlx5_tracer_handle_string_trace(struct mlx5_fw_tracer *tracer,
struct tracer_event *tracer_event) struct tracer_event *tracer_event)
{ {
...@@ -589,7 +616,7 @@ static int mlx5_tracer_handle_string_trace(struct mlx5_fw_tracer *tracer, ...@@ -589,7 +616,7 @@ static int mlx5_tracer_handle_string_trace(struct mlx5_fw_tracer *tracer,
if (tracer_event->string_event.tdsn == 0) { if (tracer_event->string_event.tdsn == 0) {
cur_string = mlx5_tracer_get_string(tracer, tracer_event); cur_string = mlx5_tracer_get_string(tracer, tracer_event);
if (!cur_string) if (!cur_string)
return -1; return mlx5_tracer_handle_raw_string(tracer, tracer_event);
cur_string->num_of_params = mlx5_tracer_get_num_of_params(cur_string->string); cur_string->num_of_params = mlx5_tracer_get_num_of_params(cur_string->string);
cur_string->last_param_num = 0; cur_string->last_param_num = 0;
...@@ -602,9 +629,9 @@ static int mlx5_tracer_handle_string_trace(struct mlx5_fw_tracer *tracer, ...@@ -602,9 +629,9 @@ static int mlx5_tracer_handle_string_trace(struct mlx5_fw_tracer *tracer,
} else { } else {
cur_string = mlx5_tracer_message_get(tracer, tracer_event); cur_string = mlx5_tracer_message_get(tracer, tracer_event);
if (!cur_string) { if (!cur_string) {
pr_debug("%s Got string event for unknown string tdsm: %d\n", pr_debug("%s Got string event for unknown string tmsn: %d\n",
__func__, tracer_event->string_event.tmsn); __func__, tracer_event->string_event.tmsn);
return -1; return mlx5_tracer_handle_raw_string(tracer, tracer_event);
} }
cur_string->last_param_num += 1; cur_string->last_param_num += 1;
if (cur_string->last_param_num > TRACER_MAX_PARAMS) { if (cur_string->last_param_num > TRACER_MAX_PARAMS) {
...@@ -930,6 +957,14 @@ int mlx5_fw_tracer_get_saved_traces_objects(struct mlx5_fw_tracer *tracer, ...@@ -930,6 +957,14 @@ int mlx5_fw_tracer_get_saved_traces_objects(struct mlx5_fw_tracer *tracer,
return err; return err;
} }
static void mlx5_fw_tracer_update_db(struct work_struct *work)
{
struct mlx5_fw_tracer *tracer =
container_of(work, struct mlx5_fw_tracer, update_db_work);
mlx5_fw_tracer_reload(tracer);
}
/* Create software resources (Buffers, etc ..) */ /* Create software resources (Buffers, etc ..) */
struct mlx5_fw_tracer *mlx5_fw_tracer_create(struct mlx5_core_dev *dev) struct mlx5_fw_tracer *mlx5_fw_tracer_create(struct mlx5_core_dev *dev)
{ {
...@@ -957,6 +992,8 @@ struct mlx5_fw_tracer *mlx5_fw_tracer_create(struct mlx5_core_dev *dev) ...@@ -957,6 +992,8 @@ struct mlx5_fw_tracer *mlx5_fw_tracer_create(struct mlx5_core_dev *dev)
INIT_WORK(&tracer->ownership_change_work, mlx5_fw_tracer_ownership_change); INIT_WORK(&tracer->ownership_change_work, mlx5_fw_tracer_ownership_change);
INIT_WORK(&tracer->read_fw_strings_work, mlx5_tracer_read_strings_db); INIT_WORK(&tracer->read_fw_strings_work, mlx5_tracer_read_strings_db);
INIT_WORK(&tracer->handle_traces_work, mlx5_fw_tracer_handle_traces); INIT_WORK(&tracer->handle_traces_work, mlx5_fw_tracer_handle_traces);
INIT_WORK(&tracer->update_db_work, mlx5_fw_tracer_update_db);
mutex_init(&tracer->state_lock);
err = mlx5_query_mtrc_caps(tracer); err = mlx5_query_mtrc_caps(tracer);
...@@ -1003,11 +1040,15 @@ int mlx5_fw_tracer_init(struct mlx5_fw_tracer *tracer) ...@@ -1003,11 +1040,15 @@ int mlx5_fw_tracer_init(struct mlx5_fw_tracer *tracer)
if (IS_ERR_OR_NULL(tracer)) if (IS_ERR_OR_NULL(tracer))
return 0; return 0;
dev = tracer->dev;
if (!tracer->str_db.loaded) if (!tracer->str_db.loaded)
queue_work(tracer->work_queue, &tracer->read_fw_strings_work); queue_work(tracer->work_queue, &tracer->read_fw_strings_work);
mutex_lock(&tracer->state_lock);
if (test_and_set_bit(MLX5_TRACER_STATE_UP, &tracer->state))
goto unlock;
dev = tracer->dev;
err = mlx5_core_alloc_pd(dev, &tracer->buff.pdn); err = mlx5_core_alloc_pd(dev, &tracer->buff.pdn);
if (err) { if (err) {
mlx5_core_warn(dev, "FWTracer: Failed to allocate PD %d\n", err); mlx5_core_warn(dev, "FWTracer: Failed to allocate PD %d\n", err);
...@@ -1028,6 +1069,8 @@ int mlx5_fw_tracer_init(struct mlx5_fw_tracer *tracer) ...@@ -1028,6 +1069,8 @@ int mlx5_fw_tracer_init(struct mlx5_fw_tracer *tracer)
mlx5_core_warn(dev, "FWTracer: Failed to start tracer %d\n", err); mlx5_core_warn(dev, "FWTracer: Failed to start tracer %d\n", err);
goto err_notifier_unregister; goto err_notifier_unregister;
} }
unlock:
mutex_unlock(&tracer->state_lock);
return 0; return 0;
err_notifier_unregister: err_notifier_unregister:
...@@ -1037,6 +1080,7 @@ int mlx5_fw_tracer_init(struct mlx5_fw_tracer *tracer) ...@@ -1037,6 +1080,7 @@ int mlx5_fw_tracer_init(struct mlx5_fw_tracer *tracer)
mlx5_core_dealloc_pd(dev, tracer->buff.pdn); mlx5_core_dealloc_pd(dev, tracer->buff.pdn);
err_cancel_work: err_cancel_work:
cancel_work_sync(&tracer->read_fw_strings_work); cancel_work_sync(&tracer->read_fw_strings_work);
mutex_unlock(&tracer->state_lock);
return err; return err;
} }
...@@ -1046,17 +1090,27 @@ void mlx5_fw_tracer_cleanup(struct mlx5_fw_tracer *tracer) ...@@ -1046,17 +1090,27 @@ void mlx5_fw_tracer_cleanup(struct mlx5_fw_tracer *tracer)
if (IS_ERR_OR_NULL(tracer)) if (IS_ERR_OR_NULL(tracer))
return; return;
mutex_lock(&tracer->state_lock);
if (!test_and_clear_bit(MLX5_TRACER_STATE_UP, &tracer->state))
goto unlock;
mlx5_core_dbg(tracer->dev, "FWTracer: Cleanup, is owner ? (%d)\n", mlx5_core_dbg(tracer->dev, "FWTracer: Cleanup, is owner ? (%d)\n",
tracer->owner); tracer->owner);
mlx5_eq_notifier_unregister(tracer->dev, &tracer->nb); mlx5_eq_notifier_unregister(tracer->dev, &tracer->nb);
cancel_work_sync(&tracer->ownership_change_work); cancel_work_sync(&tracer->ownership_change_work);
cancel_work_sync(&tracer->handle_traces_work); cancel_work_sync(&tracer->handle_traces_work);
/* It is valid to get here from update_db_work. Hence, don't wait for
* update_db_work to finished.
*/
cancel_work(&tracer->update_db_work);
if (tracer->owner) if (tracer->owner)
mlx5_fw_tracer_ownership_release(tracer); mlx5_fw_tracer_ownership_release(tracer);
mlx5_core_destroy_mkey(tracer->dev, tracer->buff.mkey); mlx5_core_destroy_mkey(tracer->dev, tracer->buff.mkey);
mlx5_core_dealloc_pd(tracer->dev, tracer->buff.pdn); mlx5_core_dealloc_pd(tracer->dev, tracer->buff.pdn);
unlock:
mutex_unlock(&tracer->state_lock);
} }
/* Free software resources (Buffers, etc ..) */ /* Free software resources (Buffers, etc ..) */
...@@ -1073,6 +1127,7 @@ void mlx5_fw_tracer_destroy(struct mlx5_fw_tracer *tracer) ...@@ -1073,6 +1127,7 @@ void mlx5_fw_tracer_destroy(struct mlx5_fw_tracer *tracer)
mlx5_fw_tracer_clean_saved_traces_array(tracer); mlx5_fw_tracer_clean_saved_traces_array(tracer);
mlx5_fw_tracer_free_strings_db(tracer); mlx5_fw_tracer_free_strings_db(tracer);
mlx5_fw_tracer_destroy_log_buf(tracer); mlx5_fw_tracer_destroy_log_buf(tracer);
mutex_destroy(&tracer->state_lock);
destroy_workqueue(tracer->work_queue); destroy_workqueue(tracer->work_queue);
kvfree(tracer); kvfree(tracer);
} }
...@@ -1082,6 +1137,8 @@ static int mlx5_fw_tracer_recreate_strings_db(struct mlx5_fw_tracer *tracer) ...@@ -1082,6 +1137,8 @@ static int mlx5_fw_tracer_recreate_strings_db(struct mlx5_fw_tracer *tracer)
struct mlx5_core_dev *dev; struct mlx5_core_dev *dev;
int err; int err;
if (test_and_set_bit(MLX5_TRACER_RECREATE_DB, &tracer->state))
return 0;
cancel_work_sync(&tracer->read_fw_strings_work); cancel_work_sync(&tracer->read_fw_strings_work);
mlx5_fw_tracer_clean_ready_list(tracer); mlx5_fw_tracer_clean_ready_list(tracer);
mlx5_fw_tracer_clean_print_hash(tracer); mlx5_fw_tracer_clean_print_hash(tracer);
...@@ -1092,17 +1149,18 @@ static int mlx5_fw_tracer_recreate_strings_db(struct mlx5_fw_tracer *tracer) ...@@ -1092,17 +1149,18 @@ static int mlx5_fw_tracer_recreate_strings_db(struct mlx5_fw_tracer *tracer)
err = mlx5_query_mtrc_caps(tracer); err = mlx5_query_mtrc_caps(tracer);
if (err) { if (err) {
mlx5_core_dbg(dev, "FWTracer: Failed to query capabilities %d\n", err); mlx5_core_dbg(dev, "FWTracer: Failed to query capabilities %d\n", err);
return err; goto out;
} }
err = mlx5_fw_tracer_allocate_strings_db(tracer); err = mlx5_fw_tracer_allocate_strings_db(tracer);
if (err) { if (err) {
mlx5_core_warn(dev, "FWTracer: Allocate strings DB failed %d\n", err); mlx5_core_warn(dev, "FWTracer: Allocate strings DB failed %d\n", err);
return err; goto out;
} }
mlx5_fw_tracer_init_saved_traces_array(tracer); mlx5_fw_tracer_init_saved_traces_array(tracer);
out:
return 0; clear_bit(MLX5_TRACER_RECREATE_DB, &tracer->state);
return err;
} }
int mlx5_fw_tracer_reload(struct mlx5_fw_tracer *tracer) int mlx5_fw_tracer_reload(struct mlx5_fw_tracer *tracer)
...@@ -1142,6 +1200,9 @@ static int fw_tracer_event(struct notifier_block *nb, unsigned long action, void ...@@ -1142,6 +1200,9 @@ static int fw_tracer_event(struct notifier_block *nb, unsigned long action, void
case MLX5_TRACER_SUBTYPE_TRACES_AVAILABLE: case MLX5_TRACER_SUBTYPE_TRACES_AVAILABLE:
queue_work(tracer->work_queue, &tracer->handle_traces_work); queue_work(tracer->work_queue, &tracer->handle_traces_work);
break; break;
case MLX5_TRACER_SUBTYPE_STRINGS_DB_UPDATE:
queue_work(tracer->work_queue, &tracer->update_db_work);
break;
default: default:
mlx5_core_dbg(dev, "FWTracer: Event with unrecognized subtype: sub_type %d\n", mlx5_core_dbg(dev, "FWTracer: Event with unrecognized subtype: sub_type %d\n",
eqe->sub_type); eqe->sub_type);
......
...@@ -63,6 +63,11 @@ struct mlx5_fw_trace_data { ...@@ -63,6 +63,11 @@ struct mlx5_fw_trace_data {
char msg[TRACE_STR_MSG]; char msg[TRACE_STR_MSG];
}; };
enum mlx5_fw_tracer_state {
MLX5_TRACER_STATE_UP = BIT(0),
MLX5_TRACER_RECREATE_DB = BIT(1),
};
struct mlx5_fw_tracer { struct mlx5_fw_tracer {
struct mlx5_core_dev *dev; struct mlx5_core_dev *dev;
struct mlx5_nb nb; struct mlx5_nb nb;
...@@ -104,6 +109,9 @@ struct mlx5_fw_tracer { ...@@ -104,6 +109,9 @@ struct mlx5_fw_tracer {
struct work_struct handle_traces_work; struct work_struct handle_traces_work;
struct hlist_head hash[MESSAGE_HASH_SIZE]; struct hlist_head hash[MESSAGE_HASH_SIZE];
struct list_head ready_strings_list; struct list_head ready_strings_list;
struct work_struct update_db_work;
struct mutex state_lock; /* Synchronize update work with reload flows */
unsigned long state;
}; };
struct tracer_string_format { struct tracer_string_format {
...@@ -158,6 +166,7 @@ struct tracer_event { ...@@ -158,6 +166,7 @@ struct tracer_event {
struct tracer_string_event string_event; struct tracer_string_event string_event;
struct tracer_timestamp_event timestamp_event; struct tracer_timestamp_event timestamp_event;
}; };
u64 *out;
}; };
struct mlx5_ifc_tracer_event_bits { struct mlx5_ifc_tracer_event_bits {
......
...@@ -314,11 +314,11 @@ static int port_set_buffer(struct mlx5e_priv *priv, ...@@ -314,11 +314,11 @@ static int port_set_buffer(struct mlx5e_priv *priv,
err = port_update_shared_buffer(priv->mdev, current_headroom_size, err = port_update_shared_buffer(priv->mdev, current_headroom_size,
new_headroom_size); new_headroom_size);
if (err) if (err)
return err; goto out;
err = port_update_pool_cfg(priv->mdev, port_buffer); err = port_update_pool_cfg(priv->mdev, port_buffer);
if (err) if (err)
return err; goto out;
err = mlx5e_port_set_pbmc(mdev, in); err = mlx5e_port_set_pbmc(mdev, in);
out: out:
......
...@@ -44,19 +44,17 @@ parse_tc_vlan_action(struct mlx5e_priv *priv, ...@@ -44,19 +44,17 @@ parse_tc_vlan_action(struct mlx5e_priv *priv,
return -EOPNOTSUPP; return -EOPNOTSUPP;
} }
switch (act->id) { if (!mlx5_eswitch_vlan_actions_supported(priv->mdev, vlan_idx)) {
case FLOW_ACTION_VLAN_POP: NL_SET_ERR_MSG_MOD(extack, "firmware vlan actions is not supported");
if (vlan_idx) {
if (!mlx5_eswitch_vlan_actions_supported(priv->mdev,
MLX5_FS_VLAN_DEPTH)) {
NL_SET_ERR_MSG_MOD(extack, "vlan pop action is not supported");
return -EOPNOTSUPP; return -EOPNOTSUPP;
} }
switch (act->id) {
case FLOW_ACTION_VLAN_POP:
if (vlan_idx)
*action |= MLX5_FLOW_CONTEXT_ACTION_VLAN_POP_2; *action |= MLX5_FLOW_CONTEXT_ACTION_VLAN_POP_2;
} else { else
*action |= MLX5_FLOW_CONTEXT_ACTION_VLAN_POP; *action |= MLX5_FLOW_CONTEXT_ACTION_VLAN_POP;
}
break; break;
case FLOW_ACTION_VLAN_PUSH: case FLOW_ACTION_VLAN_PUSH:
attr->vlan_vid[vlan_idx] = act->vlan.vid; attr->vlan_vid[vlan_idx] = act->vlan.vid;
...@@ -65,25 +63,10 @@ parse_tc_vlan_action(struct mlx5e_priv *priv, ...@@ -65,25 +63,10 @@ parse_tc_vlan_action(struct mlx5e_priv *priv,
if (!attr->vlan_proto[vlan_idx]) if (!attr->vlan_proto[vlan_idx])
attr->vlan_proto[vlan_idx] = htons(ETH_P_8021Q); attr->vlan_proto[vlan_idx] = htons(ETH_P_8021Q);
if (vlan_idx) { if (vlan_idx)
if (!mlx5_eswitch_vlan_actions_supported(priv->mdev,
MLX5_FS_VLAN_DEPTH)) {
NL_SET_ERR_MSG_MOD(extack,
"vlan push action is not supported for vlan depth > 1");
return -EOPNOTSUPP;
}
*action |= MLX5_FLOW_CONTEXT_ACTION_VLAN_PUSH_2; *action |= MLX5_FLOW_CONTEXT_ACTION_VLAN_PUSH_2;
} else { else
if (!mlx5_eswitch_vlan_actions_supported(priv->mdev, 1) &&
(act->vlan.proto != htons(ETH_P_8021Q) ||
act->vlan.prio)) {
NL_SET_ERR_MSG_MOD(extack, "vlan push action is not supported");
return -EOPNOTSUPP;
}
*action |= MLX5_FLOW_CONTEXT_ACTION_VLAN_PUSH; *action |= MLX5_FLOW_CONTEXT_ACTION_VLAN_PUSH;
}
break; break;
case FLOW_ACTION_VLAN_POP_ETH: case FLOW_ACTION_VLAN_POP_ETH:
parse_state->eth_pop = true; parse_state->eth_pop = true;
......
...@@ -92,7 +92,6 @@ static void mlx5e_ipsec_packet_setup(void *obj, u32 pdn, ...@@ -92,7 +92,6 @@ static void mlx5e_ipsec_packet_setup(void *obj, u32 pdn,
MLX5_SET(ipsec_aso, aso_ctx, remove_flow_pkt_cnt, MLX5_SET(ipsec_aso, aso_ctx, remove_flow_pkt_cnt,
lower_32_bits(attrs->hard_packet_limit)); lower_32_bits(attrs->hard_packet_limit));
MLX5_SET(ipsec_aso, aso_ctx, hard_lft_arm, 1); MLX5_SET(ipsec_aso, aso_ctx, hard_lft_arm, 1);
MLX5_SET(ipsec_aso, aso_ctx, remove_flow_enable, 1);
} }
if (attrs->soft_packet_limit != XFRM_INF) { if (attrs->soft_packet_limit != XFRM_INF) {
...@@ -329,8 +328,7 @@ static void mlx5e_ipsec_handle_event(struct work_struct *_work) ...@@ -329,8 +328,7 @@ static void mlx5e_ipsec_handle_event(struct work_struct *_work)
if (attrs->soft_packet_limit != XFRM_INF) if (attrs->soft_packet_limit != XFRM_INF)
if (!MLX5_GET(ipsec_aso, aso->ctx, soft_lft_arm) || if (!MLX5_GET(ipsec_aso, aso->ctx, soft_lft_arm) ||
!MLX5_GET(ipsec_aso, aso->ctx, hard_lft_arm) || !MLX5_GET(ipsec_aso, aso->ctx, hard_lft_arm))
!MLX5_GET(ipsec_aso, aso->ctx, remove_flow_enable))
xfrm_state_check_expire(sa_entry->x); xfrm_state_check_expire(sa_entry->x);
unlock: unlock:
......
...@@ -899,8 +899,6 @@ static void mlx5e_tls_tx_debugfs_init(struct mlx5e_tls *tls, ...@@ -899,8 +899,6 @@ static void mlx5e_tls_tx_debugfs_init(struct mlx5e_tls *tls,
return; return;
tls->debugfs.dfs_tx = debugfs_create_dir("tx", dfs_root); tls->debugfs.dfs_tx = debugfs_create_dir("tx", dfs_root);
if (!tls->debugfs.dfs_tx)
return;
debugfs_create_size_t("pool_size", 0400, tls->debugfs.dfs_tx, debugfs_create_size_t("pool_size", 0400, tls->debugfs.dfs_tx,
&tls->tx_pool->size); &tls->tx_pool->size);
......
...@@ -3002,32 +3002,37 @@ int mlx5e_safe_switch_params(struct mlx5e_priv *priv, ...@@ -3002,32 +3002,37 @@ int mlx5e_safe_switch_params(struct mlx5e_priv *priv,
mlx5e_fp_preactivate preactivate, mlx5e_fp_preactivate preactivate,
void *context, bool reset) void *context, bool reset)
{ {
struct mlx5e_channels new_chs = {}; struct mlx5e_channels *new_chs;
int err; int err;
reset &= test_bit(MLX5E_STATE_OPENED, &priv->state); reset &= test_bit(MLX5E_STATE_OPENED, &priv->state);
if (!reset) if (!reset)
return mlx5e_switch_priv_params(priv, params, preactivate, context); return mlx5e_switch_priv_params(priv, params, preactivate, context);
new_chs.params = *params; new_chs = kzalloc(sizeof(*new_chs), GFP_KERNEL);
if (!new_chs)
return -ENOMEM;
new_chs->params = *params;
mlx5e_selq_prepare_params(&priv->selq, &new_chs.params); mlx5e_selq_prepare_params(&priv->selq, &new_chs->params);
err = mlx5e_open_channels(priv, &new_chs); err = mlx5e_open_channels(priv, new_chs);
if (err) if (err)
goto err_cancel_selq; goto err_cancel_selq;
err = mlx5e_switch_priv_channels(priv, &new_chs, preactivate, context); err = mlx5e_switch_priv_channels(priv, new_chs, preactivate, context);
if (err) if (err)
goto err_close; goto err_close;
kfree(new_chs);
return 0; return 0;
err_close: err_close:
mlx5e_close_channels(&new_chs); mlx5e_close_channels(new_chs);
err_cancel_selq: err_cancel_selq:
mlx5e_selq_cancel(&priv->selq); mlx5e_selq_cancel(&priv->selq);
kfree(new_chs);
return err; return err;
} }
......
...@@ -1124,8 +1124,6 @@ static void mlx5e_tc_debugfs_init(struct mlx5e_tc_table *tc, ...@@ -1124,8 +1124,6 @@ static void mlx5e_tc_debugfs_init(struct mlx5e_tc_table *tc,
return; return;
tc->dfs_root = debugfs_create_dir("tc", dfs_root); tc->dfs_root = debugfs_create_dir("tc", dfs_root);
if (!tc->dfs_root)
return;
debugfs_create_file("hairpin_num_queues", 0644, tc->dfs_root, debugfs_create_file("hairpin_num_queues", 0644, tc->dfs_root,
&tc->hairpin_params, &fops_hairpin_queues); &tc->hairpin_params, &fops_hairpin_queues);
...@@ -1884,7 +1882,6 @@ post_process_attr(struct mlx5e_tc_flow *flow, ...@@ -1884,7 +1882,6 @@ post_process_attr(struct mlx5e_tc_flow *flow,
struct mlx5_flow_attr *attr, struct mlx5_flow_attr *attr,
struct netlink_ext_ack *extack) struct netlink_ext_ack *extack)
{ {
struct mlx5_eswitch *esw = flow->priv->mdev->priv.eswitch;
bool vf_tun; bool vf_tun;
int err = 0; int err = 0;
...@@ -1896,12 +1893,6 @@ post_process_attr(struct mlx5e_tc_flow *flow, ...@@ -1896,12 +1893,6 @@ post_process_attr(struct mlx5e_tc_flow *flow,
if (err) if (err)
goto err_out; goto err_out;
if (mlx5e_is_eswitch_flow(flow)) {
err = mlx5_eswitch_add_vlan_action(esw, attr);
if (err)
goto err_out;
}
if (attr->action & MLX5_FLOW_CONTEXT_ACTION_MOD_HDR) { if (attr->action & MLX5_FLOW_CONTEXT_ACTION_MOD_HDR) {
err = mlx5e_tc_attach_mod_hdr(flow->priv, flow, attr); err = mlx5e_tc_attach_mod_hdr(flow->priv, flow, attr);
if (err) if (err)
...@@ -2105,8 +2096,6 @@ static void mlx5e_tc_del_fdb_flow(struct mlx5e_priv *priv, ...@@ -2105,8 +2096,6 @@ static void mlx5e_tc_del_fdb_flow(struct mlx5e_priv *priv,
if (mlx5_flow_has_geneve_opt(flow)) if (mlx5_flow_has_geneve_opt(flow))
mlx5_geneve_tlv_option_del(priv->mdev->geneve); mlx5_geneve_tlv_option_del(priv->mdev->geneve);
mlx5_eswitch_del_vlan_action(esw, attr);
if (flow->decap_route) if (flow->decap_route)
mlx5e_detach_decap_route(priv, flow); mlx5e_detach_decap_route(priv, flow);
......
...@@ -222,7 +222,6 @@ struct mlx5_eswitch_fdb { ...@@ -222,7 +222,6 @@ struct mlx5_eswitch_fdb {
struct mlx5_flow_handle **send_to_vport_meta_rules; struct mlx5_flow_handle **send_to_vport_meta_rules;
struct mlx5_flow_handle *miss_rule_uni; struct mlx5_flow_handle *miss_rule_uni;
struct mlx5_flow_handle *miss_rule_multi; struct mlx5_flow_handle *miss_rule_multi;
int vlan_push_pop_refcount;
struct mlx5_fs_chains *esw_chains_priv; struct mlx5_fs_chains *esw_chains_priv;
struct { struct {
...@@ -520,10 +519,6 @@ int mlx5_devlink_port_fn_migratable_set(struct devlink_port *port, bool enable, ...@@ -520,10 +519,6 @@ int mlx5_devlink_port_fn_migratable_set(struct devlink_port *port, bool enable,
struct netlink_ext_ack *extack); struct netlink_ext_ack *extack);
void *mlx5_eswitch_get_uplink_priv(struct mlx5_eswitch *esw, u8 rep_type); void *mlx5_eswitch_get_uplink_priv(struct mlx5_eswitch *esw, u8 rep_type);
int mlx5_eswitch_add_vlan_action(struct mlx5_eswitch *esw,
struct mlx5_flow_attr *attr);
int mlx5_eswitch_del_vlan_action(struct mlx5_eswitch *esw,
struct mlx5_flow_attr *attr);
int __mlx5_eswitch_set_vport_vlan(struct mlx5_eswitch *esw, int __mlx5_eswitch_set_vport_vlan(struct mlx5_eswitch *esw,
u16 vport, u16 vlan, u8 qos, u8 set_flags); u16 vport, u16 vlan, u8 qos, u8 set_flags);
......
...@@ -579,16 +579,16 @@ mlx5_eswitch_add_offloaded_rule(struct mlx5_eswitch *esw, ...@@ -579,16 +579,16 @@ mlx5_eswitch_add_offloaded_rule(struct mlx5_eswitch *esw,
if (esw->mode != MLX5_ESWITCH_OFFLOADS) if (esw->mode != MLX5_ESWITCH_OFFLOADS)
return ERR_PTR(-EOPNOTSUPP); return ERR_PTR(-EOPNOTSUPP);
if (!mlx5_eswitch_vlan_actions_supported(esw->dev, 1))
return ERR_PTR(-EOPNOTSUPP);
dest = kcalloc(MLX5_MAX_FLOW_FWD_VPORTS + 1, sizeof(*dest), GFP_KERNEL); dest = kcalloc(MLX5_MAX_FLOW_FWD_VPORTS + 1, sizeof(*dest), GFP_KERNEL);
if (!dest) if (!dest)
return ERR_PTR(-ENOMEM); return ERR_PTR(-ENOMEM);
flow_act.action = attr->action; flow_act.action = attr->action;
/* if per flow vlan pop/push is emulated, don't set that into the firmware */
if (!mlx5_eswitch_vlan_actions_supported(esw->dev, 1)) if (flow_act.action & MLX5_FLOW_CONTEXT_ACTION_VLAN_PUSH) {
flow_act.action &= ~(MLX5_FLOW_CONTEXT_ACTION_VLAN_PUSH |
MLX5_FLOW_CONTEXT_ACTION_VLAN_POP);
else if (flow_act.action & MLX5_FLOW_CONTEXT_ACTION_VLAN_PUSH) {
flow_act.vlan[0].ethtype = ntohs(esw_attr->vlan_proto[0]); flow_act.vlan[0].ethtype = ntohs(esw_attr->vlan_proto[0]);
flow_act.vlan[0].vid = esw_attr->vlan_vid[0]; flow_act.vlan[0].vid = esw_attr->vlan_vid[0];
flow_act.vlan[0].prio = esw_attr->vlan_prio[0]; flow_act.vlan[0].prio = esw_attr->vlan_prio[0];
...@@ -829,204 +829,6 @@ mlx5_eswitch_del_fwd_rule(struct mlx5_eswitch *esw, ...@@ -829,204 +829,6 @@ mlx5_eswitch_del_fwd_rule(struct mlx5_eswitch *esw,
__mlx5_eswitch_del_rule(esw, rule, attr, true); __mlx5_eswitch_del_rule(esw, rule, attr, true);
} }
static int esw_set_global_vlan_pop(struct mlx5_eswitch *esw, u8 val)
{
struct mlx5_eswitch_rep *rep;
unsigned long i;
int err = 0;
esw_debug(esw->dev, "%s applying global %s policy\n", __func__, val ? "pop" : "none");
mlx5_esw_for_each_host_func_vport(esw, i, rep, esw->esw_funcs.num_vfs) {
if (atomic_read(&rep->rep_data[REP_ETH].state) != REP_LOADED)
continue;
err = __mlx5_eswitch_set_vport_vlan(esw, rep->vport, 0, 0, val);
if (err)
goto out;
}
out:
return err;
}
static struct mlx5_eswitch_rep *
esw_vlan_action_get_vport(struct mlx5_esw_flow_attr *attr, bool push, bool pop)
{
struct mlx5_eswitch_rep *in_rep, *out_rep, *vport = NULL;
in_rep = attr->in_rep;
out_rep = attr->dests[0].rep;
if (push)
vport = in_rep;
else if (pop)
vport = out_rep;
else
vport = in_rep;
return vport;
}
static int esw_add_vlan_action_check(struct mlx5_esw_flow_attr *attr,
bool push, bool pop, bool fwd)
{
struct mlx5_eswitch_rep *in_rep, *out_rep;
if ((push || pop) && !fwd)
goto out_notsupp;
in_rep = attr->in_rep;
out_rep = attr->dests[0].rep;
if (push && in_rep->vport == MLX5_VPORT_UPLINK)
goto out_notsupp;
if (pop && out_rep->vport == MLX5_VPORT_UPLINK)
goto out_notsupp;
/* vport has vlan push configured, can't offload VF --> wire rules w.o it */
if (!push && !pop && fwd)
if (in_rep->vlan && out_rep->vport == MLX5_VPORT_UPLINK)
goto out_notsupp;
/* protects against (1) setting rules with different vlans to push and
* (2) setting rules w.o vlans (attr->vlan = 0) && w. vlans to push (!= 0)
*/
if (push && in_rep->vlan_refcount && (in_rep->vlan != attr->vlan_vid[0]))
goto out_notsupp;
return 0;
out_notsupp:
return -EOPNOTSUPP;
}
int mlx5_eswitch_add_vlan_action(struct mlx5_eswitch *esw,
struct mlx5_flow_attr *attr)
{
struct offloads_fdb *offloads = &esw->fdb_table.offloads;
struct mlx5_esw_flow_attr *esw_attr = attr->esw_attr;
struct mlx5_eswitch_rep *vport = NULL;
bool push, pop, fwd;
int err = 0;
/* nop if we're on the vlan push/pop non emulation mode */
if (mlx5_eswitch_vlan_actions_supported(esw->dev, 1))
return 0;
push = !!(attr->action & MLX5_FLOW_CONTEXT_ACTION_VLAN_PUSH);
pop = !!(attr->action & MLX5_FLOW_CONTEXT_ACTION_VLAN_POP);
fwd = !!((attr->action & MLX5_FLOW_CONTEXT_ACTION_FWD_DEST) &&
!attr->dest_chain);
mutex_lock(&esw->state_lock);
err = esw_add_vlan_action_check(esw_attr, push, pop, fwd);
if (err)
goto unlock;
attr->flags &= ~MLX5_ATTR_FLAG_VLAN_HANDLED;
vport = esw_vlan_action_get_vport(esw_attr, push, pop);
if (!push && !pop && fwd) {
/* tracks VF --> wire rules without vlan push action */
if (esw_attr->dests[0].rep->vport == MLX5_VPORT_UPLINK) {
vport->vlan_refcount++;
attr->flags |= MLX5_ATTR_FLAG_VLAN_HANDLED;
}
goto unlock;
}
if (!push && !pop)
goto unlock;
if (!(offloads->vlan_push_pop_refcount)) {
/* it's the 1st vlan rule, apply global vlan pop policy */
err = esw_set_global_vlan_pop(esw, SET_VLAN_STRIP);
if (err)
goto out;
}
offloads->vlan_push_pop_refcount++;
if (push) {
if (vport->vlan_refcount)
goto skip_set_push;
err = __mlx5_eswitch_set_vport_vlan(esw, vport->vport, esw_attr->vlan_vid[0],
0, SET_VLAN_INSERT | SET_VLAN_STRIP);
if (err)
goto out;
vport->vlan = esw_attr->vlan_vid[0];
skip_set_push:
vport->vlan_refcount++;
}
out:
if (!err)
attr->flags |= MLX5_ATTR_FLAG_VLAN_HANDLED;
unlock:
mutex_unlock(&esw->state_lock);
return err;
}
int mlx5_eswitch_del_vlan_action(struct mlx5_eswitch *esw,
struct mlx5_flow_attr *attr)
{
struct offloads_fdb *offloads = &esw->fdb_table.offloads;
struct mlx5_esw_flow_attr *esw_attr = attr->esw_attr;
struct mlx5_eswitch_rep *vport = NULL;
bool push, pop, fwd;
int err = 0;
/* nop if we're on the vlan push/pop non emulation mode */
if (mlx5_eswitch_vlan_actions_supported(esw->dev, 1))
return 0;
if (!(attr->flags & MLX5_ATTR_FLAG_VLAN_HANDLED))
return 0;
push = !!(attr->action & MLX5_FLOW_CONTEXT_ACTION_VLAN_PUSH);
pop = !!(attr->action & MLX5_FLOW_CONTEXT_ACTION_VLAN_POP);
fwd = !!(attr->action & MLX5_FLOW_CONTEXT_ACTION_FWD_DEST);
mutex_lock(&esw->state_lock);
vport = esw_vlan_action_get_vport(esw_attr, push, pop);
if (!push && !pop && fwd) {
/* tracks VF --> wire rules without vlan push action */
if (esw_attr->dests[0].rep->vport == MLX5_VPORT_UPLINK)
vport->vlan_refcount--;
goto out;
}
if (push) {
vport->vlan_refcount--;
if (vport->vlan_refcount)
goto skip_unset_push;
vport->vlan = 0;
err = __mlx5_eswitch_set_vport_vlan(esw, vport->vport,
0, 0, SET_VLAN_STRIP);
if (err)
goto out;
}
skip_unset_push:
offloads->vlan_push_pop_refcount--;
if (offloads->vlan_push_pop_refcount)
goto out;
/* no more vlan rules, stop global vlan pop policy */
err = esw_set_global_vlan_pop(esw, 0);
out:
mutex_unlock(&esw->state_lock);
return err;
}
struct mlx5_flow_handle * struct mlx5_flow_handle *
mlx5_eswitch_add_send_to_vport_rule(struct mlx5_eswitch *on_esw, mlx5_eswitch_add_send_to_vport_rule(struct mlx5_eswitch *on_esw,
struct mlx5_eswitch *from_esw, struct mlx5_eswitch *from_esw,
......
...@@ -272,8 +272,6 @@ static int mlx5_cmd_create_flow_table(struct mlx5_flow_root_namespace *ns, ...@@ -272,8 +272,6 @@ static int mlx5_cmd_create_flow_table(struct mlx5_flow_root_namespace *ns,
unsigned int size; unsigned int size;
int err; int err;
if (ft_attr->max_fte != POOL_NEXT_SIZE)
size = roundup_pow_of_two(ft_attr->max_fte);
size = mlx5_ft_pool_get_avail_sz(dev, ft->type, ft_attr->max_fte); size = mlx5_ft_pool_get_avail_sz(dev, ft->type, ft_attr->max_fte);
if (!size) if (!size)
return -ENOSPC; return -ENOSPC;
...@@ -412,11 +410,6 @@ static int mlx5_cmd_create_flow_group(struct mlx5_flow_root_namespace *ns, ...@@ -412,11 +410,6 @@ static int mlx5_cmd_create_flow_group(struct mlx5_flow_root_namespace *ns,
MLX5_CMD_OP_CREATE_FLOW_GROUP); MLX5_CMD_OP_CREATE_FLOW_GROUP);
MLX5_SET(create_flow_group_in, in, table_type, ft->type); MLX5_SET(create_flow_group_in, in, table_type, ft->type);
MLX5_SET(create_flow_group_in, in, table_id, ft->id); MLX5_SET(create_flow_group_in, in, table_id, ft->id);
if (ft->vport) {
MLX5_SET(create_flow_group_in, in, vport_number, ft->vport);
MLX5_SET(create_flow_group_in, in, other_vport, 1);
}
MLX5_SET(create_flow_group_in, in, vport_number, ft->vport); MLX5_SET(create_flow_group_in, in, vport_number, ft->vport);
MLX5_SET(create_flow_group_in, in, other_vport, MLX5_SET(create_flow_group_in, in, other_vport,
!!(ft->flags & MLX5_FLOW_TABLE_OTHER_VPORT)); !!(ft->flags & MLX5_FLOW_TABLE_OTHER_VPORT));
......
...@@ -1776,7 +1776,6 @@ static int build_match_list(struct match_list *match_head, ...@@ -1776,7 +1776,6 @@ static int build_match_list(struct match_list *match_head,
{ {
struct rhlist_head *tmp, *list; struct rhlist_head *tmp, *list;
struct mlx5_flow_group *g; struct mlx5_flow_group *g;
int err = 0;
rcu_read_lock(); rcu_read_lock();
INIT_LIST_HEAD(&match_head->list); INIT_LIST_HEAD(&match_head->list);
...@@ -1802,7 +1801,7 @@ static int build_match_list(struct match_list *match_head, ...@@ -1802,7 +1801,7 @@ static int build_match_list(struct match_list *match_head,
list_add_tail(&curr_match->list, &match_head->list); list_add_tail(&curr_match->list, &match_head->list);
} }
rcu_read_unlock(); rcu_read_unlock();
return err; return 0;
} }
static u64 matched_fgs_get_version(struct list_head *match_head) static u64 matched_fgs_get_version(struct list_head *match_head)
......
...@@ -371,6 +371,7 @@ static int mlx5_pci_link_toggle(struct mlx5_core_dev *dev) ...@@ -371,6 +371,7 @@ static int mlx5_pci_link_toggle(struct mlx5_core_dev *dev)
mlx5_core_err(dev, "PCI link not ready (0x%04x) after %llu ms\n", mlx5_core_err(dev, "PCI link not ready (0x%04x) after %llu ms\n",
reg16, mlx5_tout_ms(dev, PCI_TOGGLE)); reg16, mlx5_tout_ms(dev, PCI_TOGGLE));
err = -ETIMEDOUT; err = -ETIMEDOUT;
goto restore;
} }
do { do {
......
...@@ -62,7 +62,7 @@ enum { ...@@ -62,7 +62,7 @@ enum {
}; };
enum { enum {
MLX5_DROP_NEW_HEALTH_WORK, MLX5_DROP_HEALTH_WORK,
}; };
enum { enum {
...@@ -675,7 +675,7 @@ static void mlx5_fw_fatal_reporter_err_work(struct work_struct *work) ...@@ -675,7 +675,7 @@ static void mlx5_fw_fatal_reporter_err_work(struct work_struct *work)
devlink = priv_to_devlink(dev); devlink = priv_to_devlink(dev);
mutex_lock(&dev->intf_state_mutex); mutex_lock(&dev->intf_state_mutex);
if (test_bit(MLX5_DROP_NEW_HEALTH_WORK, &health->flags)) { if (test_bit(MLX5_DROP_HEALTH_WORK, &health->flags)) {
mlx5_core_err(dev, "health works are not permitted at this stage\n"); mlx5_core_err(dev, "health works are not permitted at this stage\n");
mutex_unlock(&dev->intf_state_mutex); mutex_unlock(&dev->intf_state_mutex);
return; return;
...@@ -771,14 +771,8 @@ static unsigned long get_next_poll_jiffies(struct mlx5_core_dev *dev) ...@@ -771,14 +771,8 @@ static unsigned long get_next_poll_jiffies(struct mlx5_core_dev *dev)
void mlx5_trigger_health_work(struct mlx5_core_dev *dev) void mlx5_trigger_health_work(struct mlx5_core_dev *dev)
{ {
struct mlx5_core_health *health = &dev->priv.health; struct mlx5_core_health *health = &dev->priv.health;
unsigned long flags;
spin_lock_irqsave(&health->wq_lock, flags);
if (!test_bit(MLX5_DROP_NEW_HEALTH_WORK, &health->flags))
queue_work(health->wq, &health->fatal_report_work); queue_work(health->wq, &health->fatal_report_work);
else
mlx5_core_err(dev, "new health works are not permitted at this stage\n");
spin_unlock_irqrestore(&health->wq_lock, flags);
} }
#define MLX5_MSEC_PER_HOUR (MSEC_PER_SEC * 60 * 60) #define MLX5_MSEC_PER_HOUR (MSEC_PER_SEC * 60 * 60)
...@@ -858,7 +852,7 @@ void mlx5_start_health_poll(struct mlx5_core_dev *dev) ...@@ -858,7 +852,7 @@ void mlx5_start_health_poll(struct mlx5_core_dev *dev)
timer_setup(&health->timer, poll_health, 0); timer_setup(&health->timer, poll_health, 0);
health->fatal_error = MLX5_SENSOR_NO_ERR; health->fatal_error = MLX5_SENSOR_NO_ERR;
clear_bit(MLX5_DROP_NEW_HEALTH_WORK, &health->flags); clear_bit(MLX5_DROP_HEALTH_WORK, &health->flags);
health->health = &dev->iseg->health; health->health = &dev->iseg->health;
health->health_counter = &dev->iseg->health_counter; health->health_counter = &dev->iseg->health_counter;
...@@ -869,13 +863,9 @@ void mlx5_start_health_poll(struct mlx5_core_dev *dev) ...@@ -869,13 +863,9 @@ void mlx5_start_health_poll(struct mlx5_core_dev *dev)
void mlx5_stop_health_poll(struct mlx5_core_dev *dev, bool disable_health) void mlx5_stop_health_poll(struct mlx5_core_dev *dev, bool disable_health)
{ {
struct mlx5_core_health *health = &dev->priv.health; struct mlx5_core_health *health = &dev->priv.health;
unsigned long flags;
if (disable_health) { if (disable_health)
spin_lock_irqsave(&health->wq_lock, flags); set_bit(MLX5_DROP_HEALTH_WORK, &health->flags);
set_bit(MLX5_DROP_NEW_HEALTH_WORK, &health->flags);
spin_unlock_irqrestore(&health->wq_lock, flags);
}
del_timer_sync(&health->timer); del_timer_sync(&health->timer);
} }
...@@ -891,11 +881,8 @@ void mlx5_start_health_fw_log_up(struct mlx5_core_dev *dev) ...@@ -891,11 +881,8 @@ void mlx5_start_health_fw_log_up(struct mlx5_core_dev *dev)
void mlx5_drain_health_wq(struct mlx5_core_dev *dev) void mlx5_drain_health_wq(struct mlx5_core_dev *dev)
{ {
struct mlx5_core_health *health = &dev->priv.health; struct mlx5_core_health *health = &dev->priv.health;
unsigned long flags;
spin_lock_irqsave(&health->wq_lock, flags); set_bit(MLX5_DROP_HEALTH_WORK, &health->flags);
set_bit(MLX5_DROP_NEW_HEALTH_WORK, &health->flags);
spin_unlock_irqrestore(&health->wq_lock, flags);
cancel_delayed_work_sync(&health->update_fw_log_ts_work); cancel_delayed_work_sync(&health->update_fw_log_ts_work);
cancel_work_sync(&health->report_work); cancel_work_sync(&health->report_work);
cancel_work_sync(&health->fatal_report_work); cancel_work_sync(&health->fatal_report_work);
...@@ -928,7 +915,6 @@ int mlx5_health_init(struct mlx5_core_dev *dev) ...@@ -928,7 +915,6 @@ int mlx5_health_init(struct mlx5_core_dev *dev)
kfree(name); kfree(name);
if (!health->wq) if (!health->wq)
goto out_err; goto out_err;
spin_lock_init(&health->wq_lock);
INIT_WORK(&health->fatal_report_work, mlx5_fw_fatal_reporter_err_work); INIT_WORK(&health->fatal_report_work, mlx5_fw_fatal_reporter_err_work);
INIT_WORK(&health->report_work, mlx5_fw_reporter_err_work); INIT_WORK(&health->report_work, mlx5_fw_reporter_err_work);
INIT_DELAYED_WORK(&health->update_fw_log_ts_work, mlx5_health_log_ts_update); INIT_DELAYED_WORK(&health->update_fw_log_ts_work, mlx5_health_log_ts_update);
......
...@@ -367,6 +367,7 @@ enum mlx5_driver_event { ...@@ -367,6 +367,7 @@ enum mlx5_driver_event {
enum { enum {
MLX5_TRACER_SUBTYPE_OWNERSHIP_CHANGE = 0x0, MLX5_TRACER_SUBTYPE_OWNERSHIP_CHANGE = 0x0,
MLX5_TRACER_SUBTYPE_TRACES_AVAILABLE = 0x1, MLX5_TRACER_SUBTYPE_TRACES_AVAILABLE = 0x1,
MLX5_TRACER_SUBTYPE_STRINGS_DB_UPDATE = 0x2,
}; };
enum { enum {
......
...@@ -430,8 +430,6 @@ struct mlx5_core_health { ...@@ -430,8 +430,6 @@ struct mlx5_core_health {
u8 synd; u8 synd;
u32 fatal_error; u32 fatal_error;
u32 crdump_size; u32 crdump_size;
/* wq spinlock to synchronize draining */
spinlock_t wq_lock;
struct workqueue_struct *wq; struct workqueue_struct *wq;
unsigned long flags; unsigned long flags;
struct work_struct fatal_report_work; struct work_struct fatal_report_work;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment