Commit 3194d6ad authored by David S. Miller's avatar David S. Miller
Browse files


Santosh Shilimkar says:

====================
rds fixes

Few rds fixes which makes rds rdma transport reliably working on mainline

First two fixes are applicable to v4.11+ stable versions and last
three patches applies to only v5.1 stable and current mainline.

Patchset is re-based against 'net' and also available on below tree
====================

Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parents 114a5c32 dc205a8d
Loading
Loading
Loading
Loading
+1 −0
Original line number Original line Diff line number Diff line
@@ -736,6 +736,7 @@ static int rds_conn_info_visitor(struct rds_conn_path *cp, void *buffer)
	cinfo->next_rx_seq = cp->cp_next_rx_seq;
	cinfo->next_rx_seq = cp->cp_next_rx_seq;
	cinfo->laddr = conn->c_laddr.s6_addr32[3];
	cinfo->laddr = conn->c_laddr.s6_addr32[3];
	cinfo->faddr = conn->c_faddr.s6_addr32[3];
	cinfo->faddr = conn->c_faddr.s6_addr32[3];
	cinfo->tos = conn->c_tos;
	strncpy(cinfo->transport, conn->c_trans->t_name,
	strncpy(cinfo->transport, conn->c_trans->t_name,
		sizeof(cinfo->transport));
		sizeof(cinfo->transport));
	cinfo->flags = 0;
	cinfo->flags = 0;
+1 −3
Original line number Original line Diff line number Diff line
@@ -15,8 +15,7 @@


#define RDS_IB_DEFAULT_RECV_WR		1024
#define RDS_IB_DEFAULT_RECV_WR		1024
#define RDS_IB_DEFAULT_SEND_WR		256
#define RDS_IB_DEFAULT_SEND_WR		256
#define RDS_IB_DEFAULT_FR_WR		256
#define RDS_IB_DEFAULT_FR_WR		512
#define RDS_IB_DEFAULT_FR_INV_WR	256


#define RDS_IB_DEFAULT_RETRY_COUNT	1
#define RDS_IB_DEFAULT_RETRY_COUNT	1


@@ -157,7 +156,6 @@ struct rds_ib_connection {


	/* To control the number of wrs from fastreg */
	/* To control the number of wrs from fastreg */
	atomic_t		i_fastreg_wrs;
	atomic_t		i_fastreg_wrs;
	atomic_t		i_fastunreg_wrs;


	/* interrupt handling */
	/* interrupt handling */
	struct tasklet_struct	i_send_tasklet;
	struct tasklet_struct	i_send_tasklet;
+2 −7
Original line number Original line Diff line number Diff line
@@ -460,10 +460,7 @@ static int rds_ib_setup_qp(struct rds_connection *conn)
	 * completion queue and send queue. This extra space is used for FRMR
	 * completion queue and send queue. This extra space is used for FRMR
	 * registration and invalidation work requests
	 * registration and invalidation work requests
	 */
	 */
	fr_queue_space = rds_ibdev->use_fastreg ?
	fr_queue_space = (rds_ibdev->use_fastreg ? RDS_IB_DEFAULT_FR_WR : 0);
			 (RDS_IB_DEFAULT_FR_WR + 1) +
			 (RDS_IB_DEFAULT_FR_INV_WR + 1)
			 : 0;


	/* add the conn now so that connection establishment has the dev */
	/* add the conn now so that connection establishment has the dev */
	rds_ib_add_conn(rds_ibdev, conn);
	rds_ib_add_conn(rds_ibdev, conn);
@@ -530,7 +527,6 @@ static int rds_ib_setup_qp(struct rds_connection *conn)
	attr.send_cq = ic->i_send_cq;
	attr.send_cq = ic->i_send_cq;
	attr.recv_cq = ic->i_recv_cq;
	attr.recv_cq = ic->i_recv_cq;
	atomic_set(&ic->i_fastreg_wrs, RDS_IB_DEFAULT_FR_WR);
	atomic_set(&ic->i_fastreg_wrs, RDS_IB_DEFAULT_FR_WR);
	atomic_set(&ic->i_fastunreg_wrs, RDS_IB_DEFAULT_FR_INV_WR);


	/*
	/*
	 * XXX this can fail if max_*_wr is too large?  Are we supposed
	 * XXX this can fail if max_*_wr is too large?  Are we supposed
@@ -1009,8 +1005,7 @@ void rds_ib_conn_path_shutdown(struct rds_conn_path *cp)
		wait_event(rds_ib_ring_empty_wait,
		wait_event(rds_ib_ring_empty_wait,
			   rds_ib_ring_empty(&ic->i_recv_ring) &&
			   rds_ib_ring_empty(&ic->i_recv_ring) &&
			   (atomic_read(&ic->i_signaled_sends) == 0) &&
			   (atomic_read(&ic->i_signaled_sends) == 0) &&
			   (atomic_read(&ic->i_fastreg_wrs) == RDS_IB_DEFAULT_FR_WR) &&
			   (atomic_read(&ic->i_fastreg_wrs) == RDS_IB_DEFAULT_FR_WR));
			   (atomic_read(&ic->i_fastunreg_wrs) == RDS_IB_DEFAULT_FR_INV_WR));
		tasklet_kill(&ic->i_send_tasklet);
		tasklet_kill(&ic->i_send_tasklet);
		tasklet_kill(&ic->i_recv_tasklet);
		tasklet_kill(&ic->i_recv_tasklet);


+5 −6
Original line number Original line Diff line number Diff line
@@ -239,8 +239,8 @@ static int rds_ib_post_inv(struct rds_ib_mr *ibmr)
	if (frmr->fr_state != FRMR_IS_INUSE)
	if (frmr->fr_state != FRMR_IS_INUSE)
		goto out;
		goto out;


	while (atomic_dec_return(&ibmr->ic->i_fastunreg_wrs) <= 0) {
	while (atomic_dec_return(&ibmr->ic->i_fastreg_wrs) <= 0) {
		atomic_inc(&ibmr->ic->i_fastunreg_wrs);
		atomic_inc(&ibmr->ic->i_fastreg_wrs);
		cpu_relax();
		cpu_relax();
	}
	}


@@ -257,7 +257,7 @@ static int rds_ib_post_inv(struct rds_ib_mr *ibmr)
	if (unlikely(ret)) {
	if (unlikely(ret)) {
		frmr->fr_state = FRMR_IS_STALE;
		frmr->fr_state = FRMR_IS_STALE;
		frmr->fr_inv = false;
		frmr->fr_inv = false;
		atomic_inc(&ibmr->ic->i_fastunreg_wrs);
		atomic_inc(&ibmr->ic->i_fastreg_wrs);
		pr_err("RDS/IB: %s returned error(%d)\n", __func__, ret);
		pr_err("RDS/IB: %s returned error(%d)\n", __func__, ret);
		goto out;
		goto out;
	}
	}
@@ -285,10 +285,9 @@ void rds_ib_mr_cqe_handler(struct rds_ib_connection *ic, struct ib_wc *wc)
	if (frmr->fr_inv) {
	if (frmr->fr_inv) {
		frmr->fr_state = FRMR_IS_FREE;
		frmr->fr_state = FRMR_IS_FREE;
		frmr->fr_inv = false;
		frmr->fr_inv = false;
		atomic_inc(&ic->i_fastreg_wrs);
	} else {
		atomic_inc(&ic->i_fastunreg_wrs);
	}
	}

	atomic_inc(&ic->i_fastreg_wrs);
}
}


void rds_ib_unreg_frmr(struct list_head *list, unsigned int *nfreed,
void rds_ib_unreg_frmr(struct list_head *list, unsigned int *nfreed,
+13 −16
Original line number Original line Diff line number Diff line
@@ -69,6 +69,16 @@ static void rds_ib_send_complete(struct rds_message *rm,
	complete(rm, notify_status);
	complete(rm, notify_status);
}
}


static void rds_ib_send_unmap_data(struct rds_ib_connection *ic,
				   struct rm_data_op *op,
				   int wc_status)
{
	if (op->op_nents)
		ib_dma_unmap_sg(ic->i_cm_id->device,
				op->op_sg, op->op_nents,
				DMA_TO_DEVICE);
}

static void rds_ib_send_unmap_rdma(struct rds_ib_connection *ic,
static void rds_ib_send_unmap_rdma(struct rds_ib_connection *ic,
				   struct rm_rdma_op *op,
				   struct rm_rdma_op *op,
				   int wc_status)
				   int wc_status)
@@ -129,21 +139,6 @@ static void rds_ib_send_unmap_atomic(struct rds_ib_connection *ic,
		rds_ib_stats_inc(s_ib_atomic_fadd);
		rds_ib_stats_inc(s_ib_atomic_fadd);
}
}


static void rds_ib_send_unmap_data(struct rds_ib_connection *ic,
				   struct rm_data_op *op,
				   int wc_status)
{
	struct rds_message *rm = container_of(op, struct rds_message, data);

	if (op->op_nents)
		ib_dma_unmap_sg(ic->i_cm_id->device,
				op->op_sg, op->op_nents,
				DMA_TO_DEVICE);

	if (rm->rdma.op_active && rm->data.op_notify)
		rds_ib_send_unmap_rdma(ic, &rm->rdma, wc_status);
}

/*
/*
 * Unmap the resources associated with a struct send_work.
 * Unmap the resources associated with a struct send_work.
 *
 *
@@ -902,7 +897,9 @@ int rds_ib_xmit_rdma(struct rds_connection *conn, struct rm_rdma_op *op)
		send->s_queued = jiffies;
		send->s_queued = jiffies;
		send->s_op = NULL;
		send->s_op = NULL;


		nr_sig += rds_ib_set_wr_signal_state(ic, send, op->op_notify);
		if (!op->op_notify)
			nr_sig += rds_ib_set_wr_signal_state(ic, send,
							     op->op_notify);


		send->s_wr.opcode = op->op_write ? IB_WR_RDMA_WRITE : IB_WR_RDMA_READ;
		send->s_wr.opcode = op->op_write ? IB_WR_RDMA_WRITE : IB_WR_RDMA_READ;
		send->s_rdma_wr.remote_addr = remote_addr;
		send->s_rdma_wr.remote_addr = remote_addr;
Loading