ctxt->count = count;
        ctxt->direction = DMA_FROM_DEVICE;
        for (i = 0; i < count; i++) {
+               atomic_inc(&xprt->sc_dma_used);
                ctxt->sge[i].addr =
                        ib_dma_map_single(xprt->sc_cm_id->device,
                                          vec[i].iov_base, vec[i].iov_len,
 
                sge_bytes = min((size_t)bc,
                                (size_t)(vec->sge[xdr_sge_no].iov_len-sge_off));
                sge[sge_no].length = sge_bytes;
+               atomic_inc(&xprt->sc_dma_used);
                sge[sge_no].addr =
                        ib_dma_map_single(xprt->sc_cm_id->device,
                                          (void *)
        ctxt->count = 1;
 
        /* Prepare the SGE for the RPCRDMA Header */
+       atomic_inc(&rdma->sc_dma_used);
        ctxt->sge[0].addr =
                ib_dma_map_page(rdma->sc_cm_id->device,
                                page, 0, PAGE_SIZE, DMA_TO_DEVICE);
        for (sge_no = 1; byte_count && sge_no < vec->count; sge_no++) {
                sge_bytes = min_t(size_t, vec->sge[sge_no].iov_len, byte_count);
                byte_count -= sge_bytes;
+               atomic_inc(&rdma->sc_dma_used);
                ctxt->sge[sge_no].addr =
                        ib_dma_map_single(rdma->sc_cm_id->device,
                                          vec->sge[sge_no].iov_base,
 
        struct svcxprt_rdma *xprt = ctxt->xprt;
        int i;
        for (i = 0; i < ctxt->count && ctxt->sge[i].length; i++) {
+               atomic_dec(&xprt->sc_dma_used);
                ib_dma_unmap_single(xprt->sc_cm_id->device,
                                    ctxt->sge[i].addr,
                                    ctxt->sge[i].length,
        cma_xprt->sc_max_requests = svcrdma_max_requests;
        cma_xprt->sc_sq_depth = svcrdma_max_requests * RPCRDMA_SQ_DEPTH_MULT;
        atomic_set(&cma_xprt->sc_sq_count, 0);
+       atomic_set(&cma_xprt->sc_ctxt_used, 0);
 
        if (!listener) {
                int reqs = cma_xprt->sc_max_requests;
                BUG_ON(sge_no >= xprt->sc_max_sge);
                page = svc_rdma_get_page();
                ctxt->pages[sge_no] = page;
+               atomic_inc(&xprt->sc_dma_used);
                pa = ib_dma_map_page(xprt->sc_cm_id->device,
                                     page, 0, PAGE_SIZE,
                                     DMA_FROM_DEVICE);
 
        /* Warn if we leaked a resource or under-referenced */
        WARN_ON(atomic_read(&rdma->sc_ctxt_used) != 0);
+       WARN_ON(atomic_read(&rdma->sc_dma_used) != 0);
 
        /* Destroy the QP if present (not a listener) */
        if (rdma->sc_qp && !IS_ERR(rdma->sc_qp))
        length = svc_rdma_xdr_encode_error(xprt, rmsgp, err, va);
 
        /* Prepare SGE for local address */
+       atomic_inc(&xprt->sc_dma_used);
        sge.addr = ib_dma_map_page(xprt->sc_cm_id->device,
                                   p, 0, PAGE_SIZE, DMA_FROM_DEVICE);
        sge.lkey = xprt->sc_phys_mr->lkey;