Thomas Bogendoerfer f1a1bd
From: Shay Drory <shayd@nvidia.com>
Thomas Bogendoerfer f1a1bd
Date: Tue, 22 Jun 2021 14:20:16 +0300
Thomas Bogendoerfer f1a1bd
Subject: net/mlx5: Refcount mlx5_irq with integer
Thomas Bogendoerfer f1a1bd
Patch-mainline: v5.15-rc1
Thomas Bogendoerfer f1a1bd
Git-commit: 2d0b41a3767941b53160c940cdaf596a99f50fb6
Thomas Bogendoerfer f1a1bd
References: jsc#SLE-19253
Thomas Bogendoerfer f1a1bd
Thomas Bogendoerfer f1a1bd
Currently, all access to mlx5 IRQs are done undere a lock. Hance, there
Thomas Bogendoerfer f1a1bd
isn't a reason to have kref in struct mlx5_irq.
Thomas Bogendoerfer f1a1bd
Switch it to integer.
Thomas Bogendoerfer f1a1bd
Thomas Bogendoerfer f1a1bd
Signed-off-by: Shay Drory <shayd@nvidia.com>
Thomas Bogendoerfer f1a1bd
Reviewed-by: Parav Pandit <parav@nvidia.com>
Thomas Bogendoerfer f1a1bd
Signed-off-by: Saeed Mahameed <saeedm@nvidia.com>
Thomas Bogendoerfer f1a1bd
Acked-by: Thomas Bogendoerfer <tbogendoerfer@suse.de>
Thomas Bogendoerfer f1a1bd
---
Thomas Bogendoerfer f1a1bd
 drivers/net/ethernet/mellanox/mlx5/core/pci_irq.c |   65 ++++++++++++++--------
Thomas Bogendoerfer f1a1bd
 1 file changed, 44 insertions(+), 21 deletions(-)
Thomas Bogendoerfer f1a1bd
Thomas Bogendoerfer f1a1bd
--- a/drivers/net/ethernet/mellanox/mlx5/core/pci_irq.c
Thomas Bogendoerfer f1a1bd
+++ b/drivers/net/ethernet/mellanox/mlx5/core/pci_irq.c
Thomas Bogendoerfer f1a1bd
@@ -32,7 +32,7 @@ struct mlx5_irq {
Thomas Bogendoerfer f1a1bd
 	cpumask_var_t mask;
Thomas Bogendoerfer f1a1bd
 	char name[MLX5_MAX_IRQ_NAME];
Thomas Bogendoerfer f1a1bd
 	struct mlx5_irq_pool *pool;
Thomas Bogendoerfer f1a1bd
-	struct kref kref;
Thomas Bogendoerfer f1a1bd
+	int refcount;
Thomas Bogendoerfer f1a1bd
 	u32 index;
Thomas Bogendoerfer f1a1bd
 	int irqn;
Thomas Bogendoerfer f1a1bd
 };
Thomas Bogendoerfer f1a1bd
@@ -138,9 +138,8 @@ out:
Thomas Bogendoerfer f1a1bd
 	return ret;
Thomas Bogendoerfer f1a1bd
 }
Thomas Bogendoerfer f1a1bd
 
Thomas Bogendoerfer f1a1bd
-static void irq_release(struct kref *kref)
Thomas Bogendoerfer f1a1bd
+static void irq_release(struct mlx5_irq *irq)
Thomas Bogendoerfer f1a1bd
 {
Thomas Bogendoerfer f1a1bd
-	struct mlx5_irq *irq = container_of(kref, struct mlx5_irq, kref);
Thomas Bogendoerfer f1a1bd
 	struct mlx5_irq_pool *pool = irq->pool;
Thomas Bogendoerfer f1a1bd
 
Thomas Bogendoerfer f1a1bd
 	xa_erase(&pool->irqs, irq->index);
Thomas Bogendoerfer f1a1bd
@@ -159,10 +158,31 @@ static void irq_put(struct mlx5_irq *irq
Thomas Bogendoerfer f1a1bd
 	struct mlx5_irq_pool *pool = irq->pool;
Thomas Bogendoerfer f1a1bd
 
Thomas Bogendoerfer f1a1bd
 	mutex_lock(&pool->lock);
Thomas Bogendoerfer f1a1bd
-	kref_put(&irq->kref, irq_release);
Thomas Bogendoerfer f1a1bd
+	irq->refcount--;
Thomas Bogendoerfer f1a1bd
+	if (!irq->refcount)
Thomas Bogendoerfer f1a1bd
+		irq_release(irq);
Thomas Bogendoerfer f1a1bd
 	mutex_unlock(&pool->lock);
Thomas Bogendoerfer f1a1bd
 }
Thomas Bogendoerfer f1a1bd
 
Thomas Bogendoerfer f1a1bd
+static int irq_get_locked(struct mlx5_irq *irq)
Thomas Bogendoerfer f1a1bd
+{
Thomas Bogendoerfer f1a1bd
+	lockdep_assert_held(&irq->pool->lock);
Thomas Bogendoerfer f1a1bd
+	if (WARN_ON_ONCE(!irq->refcount))
Thomas Bogendoerfer f1a1bd
+		return 0;
Thomas Bogendoerfer f1a1bd
+	irq->refcount++;
Thomas Bogendoerfer f1a1bd
+	return 1;
Thomas Bogendoerfer f1a1bd
+}
Thomas Bogendoerfer f1a1bd
+
Thomas Bogendoerfer f1a1bd
+static int irq_get(struct mlx5_irq *irq)
Thomas Bogendoerfer f1a1bd
+{
Thomas Bogendoerfer f1a1bd
+	int err;
Thomas Bogendoerfer f1a1bd
+
Thomas Bogendoerfer f1a1bd
+	mutex_lock(&irq->pool->lock);
Thomas Bogendoerfer f1a1bd
+	err = irq_get_locked(irq);
Thomas Bogendoerfer f1a1bd
+	mutex_unlock(&irq->pool->lock);
Thomas Bogendoerfer f1a1bd
+	return err;
Thomas Bogendoerfer f1a1bd
+}
Thomas Bogendoerfer f1a1bd
+
Thomas Bogendoerfer f1a1bd
 static irqreturn_t irq_int_handler(int irq, void *nh)
Thomas Bogendoerfer f1a1bd
 {
Thomas Bogendoerfer f1a1bd
 	atomic_notifier_call_chain(nh, 0, NULL);
Thomas Bogendoerfer f1a1bd
@@ -215,7 +235,7 @@ static struct mlx5_irq *irq_request(stru
Thomas Bogendoerfer f1a1bd
 		goto err_cpumask;
Thomas Bogendoerfer f1a1bd
 	}
Thomas Bogendoerfer f1a1bd
 	irq->pool = pool;
Thomas Bogendoerfer f1a1bd
-	kref_init(&irq->kref);
Thomas Bogendoerfer f1a1bd
+	irq->refcount = 1;
Thomas Bogendoerfer f1a1bd
 	irq->index = i;
Thomas Bogendoerfer f1a1bd
 	err = xa_err(xa_store(&pool->irqs, irq->index, irq, GFP_KERNEL));
Thomas Bogendoerfer f1a1bd
 	if (err) {
Thomas Bogendoerfer f1a1bd
@@ -235,18 +255,18 @@ err_req_irq:
Thomas Bogendoerfer f1a1bd
 
Thomas Bogendoerfer f1a1bd
 int mlx5_irq_attach_nb(struct mlx5_irq *irq, struct notifier_block *nb)
Thomas Bogendoerfer f1a1bd
 {
Thomas Bogendoerfer f1a1bd
-	int err;
Thomas Bogendoerfer f1a1bd
+	int ret;
Thomas Bogendoerfer f1a1bd
 
Thomas Bogendoerfer f1a1bd
-	err = kref_get_unless_zero(&irq->kref);
Thomas Bogendoerfer f1a1bd
-	if (WARN_ON_ONCE(!err))
Thomas Bogendoerfer f1a1bd
+	ret = irq_get(irq);
Thomas Bogendoerfer f1a1bd
+	if (!ret)
Thomas Bogendoerfer f1a1bd
 		/* Something very bad happens here, we are enabling EQ
Thomas Bogendoerfer f1a1bd
 		 * on non-existing IRQ.
Thomas Bogendoerfer f1a1bd
 		 */
Thomas Bogendoerfer f1a1bd
 		return -ENOENT;
Thomas Bogendoerfer f1a1bd
-	err = atomic_notifier_chain_register(&irq->nh, nb);
Thomas Bogendoerfer f1a1bd
-	if (err)
Thomas Bogendoerfer f1a1bd
+	ret = atomic_notifier_chain_register(&irq->nh, nb);
Thomas Bogendoerfer f1a1bd
+	if (ret)
Thomas Bogendoerfer f1a1bd
 		irq_put(irq);
Thomas Bogendoerfer f1a1bd
-	return err;
Thomas Bogendoerfer f1a1bd
+	return ret;
Thomas Bogendoerfer f1a1bd
 }
Thomas Bogendoerfer f1a1bd
 
Thomas Bogendoerfer f1a1bd
 int mlx5_irq_detach_nb(struct mlx5_irq *irq, struct notifier_block *nb)
Thomas Bogendoerfer f1a1bd
@@ -304,10 +324,9 @@ static struct mlx5_irq *irq_pool_find_le
Thomas Bogendoerfer f1a1bd
 	xa_for_each_range(&pool->irqs, index, iter, start, end) {
Thomas Bogendoerfer f1a1bd
 		if (!cpumask_equal(iter->mask, affinity))
Thomas Bogendoerfer f1a1bd
 			continue;
Thomas Bogendoerfer f1a1bd
-		if (kref_read(&iter->kref) < pool->min_threshold)
Thomas Bogendoerfer f1a1bd
+		if (iter->refcount < pool->min_threshold)
Thomas Bogendoerfer f1a1bd
 			return iter;
Thomas Bogendoerfer f1a1bd
-		if (!irq || kref_read(&iter->kref) <
Thomas Bogendoerfer f1a1bd
-		    kref_read(&irq->kref))
Thomas Bogendoerfer f1a1bd
+		if (!irq || iter->refcount < irq->refcount)
Thomas Bogendoerfer f1a1bd
 			irq = iter;
Thomas Bogendoerfer f1a1bd
 	}
Thomas Bogendoerfer f1a1bd
 	return irq;
Thomas Bogendoerfer f1a1bd
@@ -322,7 +341,7 @@ static struct mlx5_irq *irq_pool_request
Thomas Bogendoerfer f1a1bd
 	mutex_lock(&pool->lock);
Thomas Bogendoerfer f1a1bd
 	least_loaded_irq = irq_pool_find_least_loaded(pool, affinity);
Thomas Bogendoerfer f1a1bd
 	if (least_loaded_irq &&
Thomas Bogendoerfer f1a1bd
-	    kref_read(&least_loaded_irq->kref) < pool->min_threshold)
Thomas Bogendoerfer f1a1bd
+	    least_loaded_irq->refcount < pool->min_threshold)
Thomas Bogendoerfer f1a1bd
 		goto out;
Thomas Bogendoerfer f1a1bd
 	new_irq = irq_pool_create_irq(pool, affinity);
Thomas Bogendoerfer f1a1bd
 	if (IS_ERR(new_irq)) {
Thomas Bogendoerfer f1a1bd
@@ -340,11 +359,11 @@ static struct mlx5_irq *irq_pool_request
Thomas Bogendoerfer f1a1bd
 	least_loaded_irq = new_irq;
Thomas Bogendoerfer f1a1bd
 	goto unlock;
Thomas Bogendoerfer f1a1bd
 out:
Thomas Bogendoerfer f1a1bd
-	kref_get(&least_loaded_irq->kref);
Thomas Bogendoerfer f1a1bd
-	if (kref_read(&least_loaded_irq->kref) > pool->max_threshold)
Thomas Bogendoerfer f1a1bd
+	irq_get_locked(least_loaded_irq);
Thomas Bogendoerfer f1a1bd
+	if (least_loaded_irq->refcount > pool->max_threshold)
Thomas Bogendoerfer f1a1bd
 		mlx5_core_dbg(pool->dev, "IRQ %u overloaded, pool_name: %s, %u EQs on this irq\n",
Thomas Bogendoerfer f1a1bd
 			      least_loaded_irq->irqn, pool->name,
Thomas Bogendoerfer f1a1bd
-			      kref_read(&least_loaded_irq->kref) / MLX5_EQ_REFS_PER_IRQ);
Thomas Bogendoerfer f1a1bd
+			      least_loaded_irq->refcount / MLX5_EQ_REFS_PER_IRQ);
Thomas Bogendoerfer f1a1bd
 unlock:
Thomas Bogendoerfer f1a1bd
 	mutex_unlock(&pool->lock);
Thomas Bogendoerfer f1a1bd
 	return least_loaded_irq;
Thomas Bogendoerfer f1a1bd
@@ -360,7 +379,7 @@ irq_pool_request_vector(struct mlx5_irq_
Thomas Bogendoerfer f1a1bd
 	mutex_lock(&pool->lock);
Thomas Bogendoerfer f1a1bd
 	irq = xa_load(&pool->irqs, vecidx);
Thomas Bogendoerfer f1a1bd
 	if (irq) {
Thomas Bogendoerfer f1a1bd
-		kref_get(&irq->kref);
Thomas Bogendoerfer f1a1bd
+		irq_get_locked(irq);
Thomas Bogendoerfer f1a1bd
 		goto unlock;
Thomas Bogendoerfer f1a1bd
 	}
Thomas Bogendoerfer f1a1bd
 	irq = irq_request(pool, vecidx);
Thomas Bogendoerfer f1a1bd
@@ -427,7 +446,7 @@ out:
Thomas Bogendoerfer f1a1bd
 		return irq;
Thomas Bogendoerfer f1a1bd
 	mlx5_core_dbg(dev, "irq %u mapped to cpu %*pbl, %u EQs on this irq\n",
Thomas Bogendoerfer f1a1bd
 		      irq->irqn, cpumask_pr_args(affinity),
Thomas Bogendoerfer f1a1bd
-		      kref_read(&irq->kref) / MLX5_EQ_REFS_PER_IRQ);
Thomas Bogendoerfer f1a1bd
+		      irq->refcount / MLX5_EQ_REFS_PER_IRQ);
Thomas Bogendoerfer f1a1bd
 	return irq;
Thomas Bogendoerfer f1a1bd
 }
Thomas Bogendoerfer f1a1bd
 
Thomas Bogendoerfer f1a1bd
@@ -459,8 +478,12 @@ static void irq_pool_free(struct mlx5_ir
Thomas Bogendoerfer f1a1bd
 	struct mlx5_irq *irq;
Thomas Bogendoerfer f1a1bd
 	unsigned long index;
Thomas Bogendoerfer f1a1bd
 
Thomas Bogendoerfer f1a1bd
+	/* There are cases in which we are destrying the irq_table before
Thomas Bogendoerfer f1a1bd
+	 * freeing all the IRQs, fast teardown for example. Hence, free the irqs
Thomas Bogendoerfer f1a1bd
+	 * which might not have been freed.
Thomas Bogendoerfer f1a1bd
+	 */
Thomas Bogendoerfer f1a1bd
 	xa_for_each(&pool->irqs, index, irq)
Thomas Bogendoerfer f1a1bd
-		irq_release(&irq->kref);
Thomas Bogendoerfer f1a1bd
+		irq_release(irq);
Thomas Bogendoerfer f1a1bd
 	xa_destroy(&pool->irqs);
Thomas Bogendoerfer f1a1bd
 	mutex_destroy(&pool->lock);
Thomas Bogendoerfer f1a1bd
 	kvfree(pool);