 c387e8a4ec
			
		
	
	
		c387e8a4ec
		
	
	
	
	
		
			
			Array size is MAX_PORT_GIDS, let's make sure the given index is in range. While there limit device table size to 1. Reported-by: Peter Maydell <peter.maydell@linaro.org> Signed-off-by: Yuval Shaia <yuval.shaia@oracle.com> Reviewed-by: Marcel Apfelbaum <marcel.apfelbaum@gmail.com> Message-Id: <20180430200223.4119-5-marcel.apfelbaum@gmail.com>
		
			
				
	
	
		
			701 lines
		
	
	
		
			21 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			701 lines
		
	
	
		
			21 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * QEMU paravirtual RDMA - Command channel
 | |
|  *
 | |
|  * Copyright (C) 2018 Oracle
 | |
|  * Copyright (C) 2018 Red Hat Inc
 | |
|  *
 | |
|  * Authors:
 | |
|  *     Yuval Shaia <yuval.shaia@oracle.com>
 | |
|  *     Marcel Apfelbaum <marcel@redhat.com>
 | |
|  *
 | |
|  * This work is licensed under the terms of the GNU GPL, version 2 or later.
 | |
|  * See the COPYING file in the top-level directory.
 | |
|  *
 | |
|  */
 | |
| 
 | |
| #include "qemu/osdep.h"
 | |
| #include "qemu/error-report.h"
 | |
| #include "cpu.h"
 | |
| #include <linux/types.h>
 | |
| #include "hw/hw.h"
 | |
| #include "hw/pci/pci.h"
 | |
| #include "hw/pci/pci_ids.h"
 | |
| 
 | |
| #include "../rdma_backend.h"
 | |
| #include "../rdma_rm.h"
 | |
| #include "../rdma_utils.h"
 | |
| 
 | |
| #include "pvrdma.h"
 | |
| #include "standard-headers/rdma/vmw_pvrdma-abi.h"
 | |
| 
 | |
| static void *pvrdma_map_to_pdir(PCIDevice *pdev, uint64_t pdir_dma,
 | |
|                                 uint32_t nchunks, size_t length)
 | |
| {
 | |
|     uint64_t *dir, *tbl;
 | |
|     int tbl_idx, dir_idx, addr_idx;
 | |
|     void *host_virt = NULL, *curr_page;
 | |
| 
 | |
|     if (!nchunks) {
 | |
|         pr_dbg("nchunks=0\n");
 | |
|         return NULL;
 | |
|     }
 | |
| 
 | |
|     dir = rdma_pci_dma_map(pdev, pdir_dma, TARGET_PAGE_SIZE);
 | |
|     if (!dir) {
 | |
|         error_report("PVRDMA: Failed to map to page directory");
 | |
|         return NULL;
 | |
|     }
 | |
| 
 | |
|     tbl = rdma_pci_dma_map(pdev, dir[0], TARGET_PAGE_SIZE);
 | |
|     if (!tbl) {
 | |
|         error_report("PVRDMA: Failed to map to page table 0");
 | |
|         goto out_unmap_dir;
 | |
|     }
 | |
| 
 | |
|     curr_page = rdma_pci_dma_map(pdev, (dma_addr_t)tbl[0], TARGET_PAGE_SIZE);
 | |
|     if (!curr_page) {
 | |
|         error_report("PVRDMA: Failed to map the first page");
 | |
|         goto out_unmap_tbl;
 | |
|     }
 | |
| 
 | |
|     host_virt = mremap(curr_page, 0, length, MREMAP_MAYMOVE);
 | |
|     if (host_virt == MAP_FAILED) {
 | |
|         host_virt = NULL;
 | |
|         error_report("PVRDMA: Failed to remap memory for host_virt");
 | |
|         goto out_unmap_tbl;
 | |
|     }
 | |
| 
 | |
|     rdma_pci_dma_unmap(pdev, curr_page, TARGET_PAGE_SIZE);
 | |
| 
 | |
|     pr_dbg("host_virt=%p\n", host_virt);
 | |
| 
 | |
|     dir_idx = 0;
 | |
|     tbl_idx = 1;
 | |
|     addr_idx = 1;
 | |
|     while (addr_idx < nchunks) {
 | |
|         if (tbl_idx == TARGET_PAGE_SIZE / sizeof(uint64_t)) {
 | |
|             tbl_idx = 0;
 | |
|             dir_idx++;
 | |
|             pr_dbg("Mapping to table %d\n", dir_idx);
 | |
|             rdma_pci_dma_unmap(pdev, tbl, TARGET_PAGE_SIZE);
 | |
|             tbl = rdma_pci_dma_map(pdev, dir[dir_idx], TARGET_PAGE_SIZE);
 | |
|             if (!tbl) {
 | |
|                 error_report("PVRDMA: Failed to map to page table %d", dir_idx);
 | |
|                 goto out_unmap_host_virt;
 | |
|             }
 | |
|         }
 | |
| 
 | |
|         pr_dbg("guest_dma[%d]=0x%" PRIx64 "\n", addr_idx, tbl[tbl_idx]);
 | |
| 
 | |
|         curr_page = rdma_pci_dma_map(pdev, (dma_addr_t)tbl[tbl_idx],
 | |
|                                      TARGET_PAGE_SIZE);
 | |
|         if (!curr_page) {
 | |
|             error_report("PVRDMA: Failed to map to page %d, dir %d", tbl_idx,
 | |
|                          dir_idx);
 | |
|             goto out_unmap_host_virt;
 | |
|         }
 | |
| 
 | |
|         mremap(curr_page, 0, TARGET_PAGE_SIZE, MREMAP_MAYMOVE | MREMAP_FIXED,
 | |
|                host_virt + TARGET_PAGE_SIZE * addr_idx);
 | |
| 
 | |
|         rdma_pci_dma_unmap(pdev, curr_page, TARGET_PAGE_SIZE);
 | |
| 
 | |
|         addr_idx++;
 | |
| 
 | |
|         tbl_idx++;
 | |
|     }
 | |
| 
 | |
|     goto out_unmap_tbl;
 | |
| 
 | |
| out_unmap_host_virt:
 | |
|     munmap(host_virt, length);
 | |
|     host_virt = NULL;
 | |
| 
 | |
| out_unmap_tbl:
 | |
|     rdma_pci_dma_unmap(pdev, tbl, TARGET_PAGE_SIZE);
 | |
| 
 | |
| out_unmap_dir:
 | |
|     rdma_pci_dma_unmap(pdev, dir, TARGET_PAGE_SIZE);
 | |
| 
 | |
|     return host_virt;
 | |
| }
 | |
| 
 | |
| static int query_port(PVRDMADev *dev, union pvrdma_cmd_req *req,
 | |
|                       union pvrdma_cmd_resp *rsp)
 | |
| {
 | |
|     struct pvrdma_cmd_query_port *cmd = &req->query_port;
 | |
|     struct pvrdma_cmd_query_port_resp *resp = &rsp->query_port_resp;
 | |
|     struct pvrdma_port_attr attrs = {0};
 | |
| 
 | |
|     pr_dbg("port=%d\n", cmd->port_num);
 | |
| 
 | |
|     if (rdma_backend_query_port(&dev->backend_dev,
 | |
|                                 (struct ibv_port_attr *)&attrs)) {
 | |
|         return -ENOMEM;
 | |
|     }
 | |
| 
 | |
|     memset(resp, 0, sizeof(*resp));
 | |
|     resp->hdr.response = cmd->hdr.response;
 | |
|     resp->hdr.ack = PVRDMA_CMD_QUERY_PORT_RESP;
 | |
|     resp->hdr.err = 0;
 | |
| 
 | |
|     resp->attrs.state = attrs.state;
 | |
|     resp->attrs.max_mtu = attrs.max_mtu;
 | |
|     resp->attrs.active_mtu = attrs.active_mtu;
 | |
|     resp->attrs.phys_state = attrs.phys_state;
 | |
|     resp->attrs.gid_tbl_len = MIN(MAX_PORT_GIDS, attrs.gid_tbl_len);
 | |
|     resp->attrs.max_msg_sz = 1024;
 | |
|     resp->attrs.pkey_tbl_len = MIN(MAX_PORT_PKEYS, attrs.pkey_tbl_len);
 | |
|     resp->attrs.active_width = 1;
 | |
|     resp->attrs.active_speed = 1;
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static int query_pkey(PVRDMADev *dev, union pvrdma_cmd_req *req,
 | |
|                       union pvrdma_cmd_resp *rsp)
 | |
| {
 | |
|     struct pvrdma_cmd_query_pkey *cmd = &req->query_pkey;
 | |
|     struct pvrdma_cmd_query_pkey_resp *resp = &rsp->query_pkey_resp;
 | |
| 
 | |
|     pr_dbg("port=%d\n", cmd->port_num);
 | |
|     pr_dbg("index=%d\n", cmd->index);
 | |
| 
 | |
|     memset(resp, 0, sizeof(*resp));
 | |
|     resp->hdr.response = cmd->hdr.response;
 | |
|     resp->hdr.ack = PVRDMA_CMD_QUERY_PKEY_RESP;
 | |
|     resp->hdr.err = 0;
 | |
| 
 | |
|     resp->pkey = 0x7FFF;
 | |
|     pr_dbg("pkey=0x%x\n", resp->pkey);
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static int create_pd(PVRDMADev *dev, union pvrdma_cmd_req *req,
 | |
|                      union pvrdma_cmd_resp *rsp)
 | |
| {
 | |
|     struct pvrdma_cmd_create_pd *cmd = &req->create_pd;
 | |
|     struct pvrdma_cmd_create_pd_resp *resp = &rsp->create_pd_resp;
 | |
| 
 | |
|     pr_dbg("context=0x%x\n", cmd->ctx_handle ? cmd->ctx_handle : 0);
 | |
| 
 | |
|     memset(resp, 0, sizeof(*resp));
 | |
|     resp->hdr.response = cmd->hdr.response;
 | |
|     resp->hdr.ack = PVRDMA_CMD_CREATE_PD_RESP;
 | |
|     resp->hdr.err = rdma_rm_alloc_pd(&dev->rdma_dev_res, &dev->backend_dev,
 | |
|                                      &resp->pd_handle, cmd->ctx_handle);
 | |
| 
 | |
|     pr_dbg("ret=%d\n", resp->hdr.err);
 | |
|     return resp->hdr.err;
 | |
| }
 | |
| 
 | |
| static int destroy_pd(PVRDMADev *dev, union pvrdma_cmd_req *req,
 | |
|                       union pvrdma_cmd_resp *rsp)
 | |
| {
 | |
|     struct pvrdma_cmd_destroy_pd *cmd = &req->destroy_pd;
 | |
| 
 | |
|     pr_dbg("pd_handle=%d\n", cmd->pd_handle);
 | |
| 
 | |
|     rdma_rm_dealloc_pd(&dev->rdma_dev_res, cmd->pd_handle);
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static int create_mr(PVRDMADev *dev, union pvrdma_cmd_req *req,
 | |
|                      union pvrdma_cmd_resp *rsp)
 | |
| {
 | |
|     struct pvrdma_cmd_create_mr *cmd = &req->create_mr;
 | |
|     struct pvrdma_cmd_create_mr_resp *resp = &rsp->create_mr_resp;
 | |
|     PCIDevice *pci_dev = PCI_DEVICE(dev);
 | |
|     void *host_virt = NULL;
 | |
| 
 | |
|     memset(resp, 0, sizeof(*resp));
 | |
|     resp->hdr.response = cmd->hdr.response;
 | |
|     resp->hdr.ack = PVRDMA_CMD_CREATE_MR_RESP;
 | |
| 
 | |
|     pr_dbg("pd_handle=%d\n", cmd->pd_handle);
 | |
|     pr_dbg("access_flags=0x%x\n", cmd->access_flags);
 | |
|     pr_dbg("flags=0x%x\n", cmd->flags);
 | |
| 
 | |
|     if (!(cmd->flags & PVRDMA_MR_FLAG_DMA)) {
 | |
|         host_virt = pvrdma_map_to_pdir(pci_dev, cmd->pdir_dma, cmd->nchunks,
 | |
|                                        cmd->length);
 | |
|         if (!host_virt) {
 | |
|             pr_dbg("Failed to map to pdir\n");
 | |
|             resp->hdr.err = -EINVAL;
 | |
|             goto out;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     resp->hdr.err = rdma_rm_alloc_mr(&dev->rdma_dev_res, cmd->pd_handle,
 | |
|                                      cmd->start, cmd->length, host_virt,
 | |
|                                      cmd->access_flags, &resp->mr_handle,
 | |
|                                      &resp->lkey, &resp->rkey);
 | |
|     if (host_virt && !resp->hdr.err) {
 | |
|         munmap(host_virt, cmd->length);
 | |
|     }
 | |
| 
 | |
| out:
 | |
|     pr_dbg("ret=%d\n", resp->hdr.err);
 | |
|     return resp->hdr.err;
 | |
| }
 | |
| 
 | |
| static int destroy_mr(PVRDMADev *dev, union pvrdma_cmd_req *req,
 | |
|                       union pvrdma_cmd_resp *rsp)
 | |
| {
 | |
|     struct pvrdma_cmd_destroy_mr *cmd = &req->destroy_mr;
 | |
| 
 | |
|     pr_dbg("mr_handle=%d\n", cmd->mr_handle);
 | |
| 
 | |
|     rdma_rm_dealloc_mr(&dev->rdma_dev_res, cmd->mr_handle);
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static int create_cq_ring(PCIDevice *pci_dev , PvrdmaRing **ring,
 | |
|                           uint64_t pdir_dma, uint32_t nchunks, uint32_t cqe)
 | |
| {
 | |
|     uint64_t *dir = NULL, *tbl = NULL;
 | |
|     PvrdmaRing *r;
 | |
|     int rc = -EINVAL;
 | |
|     char ring_name[MAX_RING_NAME_SZ];
 | |
| 
 | |
|     pr_dbg("pdir_dma=0x%llx\n", (long long unsigned int)pdir_dma);
 | |
|     dir = rdma_pci_dma_map(pci_dev, pdir_dma, TARGET_PAGE_SIZE);
 | |
|     if (!dir) {
 | |
|         pr_dbg("Failed to map to CQ page directory\n");
 | |
|         goto out;
 | |
|     }
 | |
| 
 | |
|     tbl = rdma_pci_dma_map(pci_dev, dir[0], TARGET_PAGE_SIZE);
 | |
|     if (!tbl) {
 | |
|         pr_dbg("Failed to map to CQ page table\n");
 | |
|         goto out;
 | |
|     }
 | |
| 
 | |
|     r = g_malloc(sizeof(*r));
 | |
|     *ring = r;
 | |
| 
 | |
|     r->ring_state = (struct pvrdma_ring *)
 | |
|         rdma_pci_dma_map(pci_dev, tbl[0], TARGET_PAGE_SIZE);
 | |
| 
 | |
|     if (!r->ring_state) {
 | |
|         pr_dbg("Failed to map to CQ ring state\n");
 | |
|         goto out_free_ring;
 | |
|     }
 | |
| 
 | |
|     sprintf(ring_name, "cq_ring_%" PRIx64, pdir_dma);
 | |
|     rc = pvrdma_ring_init(r, ring_name, pci_dev, &r->ring_state[1],
 | |
|                           cqe, sizeof(struct pvrdma_cqe),
 | |
|                           /* first page is ring state */
 | |
|                           (dma_addr_t *)&tbl[1], nchunks - 1);
 | |
|     if (rc) {
 | |
|         goto out_unmap_ring_state;
 | |
|     }
 | |
| 
 | |
|     goto out;
 | |
| 
 | |
| out_unmap_ring_state:
 | |
|     /* ring_state was in slot 1, not 0 so need to jump back */
 | |
|     rdma_pci_dma_unmap(pci_dev, --r->ring_state, TARGET_PAGE_SIZE);
 | |
| 
 | |
| out_free_ring:
 | |
|     g_free(r);
 | |
| 
 | |
| out:
 | |
|     rdma_pci_dma_unmap(pci_dev, tbl, TARGET_PAGE_SIZE);
 | |
|     rdma_pci_dma_unmap(pci_dev, dir, TARGET_PAGE_SIZE);
 | |
| 
 | |
|     return rc;
 | |
| }
 | |
| 
 | |
| static int create_cq(PVRDMADev *dev, union pvrdma_cmd_req *req,
 | |
|                      union pvrdma_cmd_resp *rsp)
 | |
| {
 | |
|     struct pvrdma_cmd_create_cq *cmd = &req->create_cq;
 | |
|     struct pvrdma_cmd_create_cq_resp *resp = &rsp->create_cq_resp;
 | |
|     PvrdmaRing *ring = NULL;
 | |
| 
 | |
|     memset(resp, 0, sizeof(*resp));
 | |
|     resp->hdr.response = cmd->hdr.response;
 | |
|     resp->hdr.ack = PVRDMA_CMD_CREATE_CQ_RESP;
 | |
| 
 | |
|     resp->cqe = cmd->cqe;
 | |
| 
 | |
|     resp->hdr.err = create_cq_ring(PCI_DEVICE(dev), &ring, cmd->pdir_dma,
 | |
|                                    cmd->nchunks, cmd->cqe);
 | |
|     if (resp->hdr.err) {
 | |
|         goto out;
 | |
|     }
 | |
| 
 | |
|     pr_dbg("ring=%p\n", ring);
 | |
| 
 | |
|     resp->hdr.err = rdma_rm_alloc_cq(&dev->rdma_dev_res, &dev->backend_dev,
 | |
|                                      cmd->cqe, &resp->cq_handle, ring);
 | |
|     resp->cqe = cmd->cqe;
 | |
| 
 | |
| out:
 | |
|     pr_dbg("ret=%d\n", resp->hdr.err);
 | |
|     return resp->hdr.err;
 | |
| }
 | |
| 
 | |
| static int destroy_cq(PVRDMADev *dev, union pvrdma_cmd_req *req,
 | |
|                       union pvrdma_cmd_resp *rsp)
 | |
| {
 | |
|     struct pvrdma_cmd_destroy_cq *cmd = &req->destroy_cq;
 | |
|     RdmaRmCQ *cq;
 | |
|     PvrdmaRing *ring;
 | |
| 
 | |
|     pr_dbg("cq_handle=%d\n", cmd->cq_handle);
 | |
| 
 | |
|     cq = rdma_rm_get_cq(&dev->rdma_dev_res, cmd->cq_handle);
 | |
|     if (!cq) {
 | |
|         pr_dbg("Invalid CQ handle\n");
 | |
|         return -EINVAL;
 | |
|     }
 | |
| 
 | |
|     ring = (PvrdmaRing *)cq->opaque;
 | |
|     pvrdma_ring_free(ring);
 | |
|     /* ring_state was in slot 1, not 0 so need to jump back */
 | |
|     rdma_pci_dma_unmap(PCI_DEVICE(dev), --ring->ring_state, TARGET_PAGE_SIZE);
 | |
|     g_free(ring);
 | |
| 
 | |
|     rdma_rm_dealloc_cq(&dev->rdma_dev_res, cmd->cq_handle);
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static int create_qp_rings(PCIDevice *pci_dev, uint64_t pdir_dma,
 | |
|                            PvrdmaRing **rings, uint32_t scqe, uint32_t smax_sge,
 | |
|                            uint32_t spages, uint32_t rcqe, uint32_t rmax_sge,
 | |
|                            uint32_t rpages)
 | |
| {
 | |
|     uint64_t *dir = NULL, *tbl = NULL;
 | |
|     PvrdmaRing *sr, *rr;
 | |
|     int rc = -EINVAL;
 | |
|     char ring_name[MAX_RING_NAME_SZ];
 | |
|     uint32_t wqe_sz;
 | |
| 
 | |
|     pr_dbg("pdir_dma=0x%llx\n", (long long unsigned int)pdir_dma);
 | |
|     dir = rdma_pci_dma_map(pci_dev, pdir_dma, TARGET_PAGE_SIZE);
 | |
|     if (!dir) {
 | |
|         pr_dbg("Failed to map to CQ page directory\n");
 | |
|         goto out;
 | |
|     }
 | |
| 
 | |
|     tbl = rdma_pci_dma_map(pci_dev, dir[0], TARGET_PAGE_SIZE);
 | |
|     if (!tbl) {
 | |
|         pr_dbg("Failed to map to CQ page table\n");
 | |
|         goto out;
 | |
|     }
 | |
| 
 | |
|     sr = g_malloc(2 * sizeof(*rr));
 | |
|     rr = &sr[1];
 | |
|     pr_dbg("sring=%p\n", sr);
 | |
|     pr_dbg("rring=%p\n", rr);
 | |
| 
 | |
|     *rings = sr;
 | |
| 
 | |
|     pr_dbg("scqe=%d\n", scqe);
 | |
|     pr_dbg("smax_sge=%d\n", smax_sge);
 | |
|     pr_dbg("spages=%d\n", spages);
 | |
|     pr_dbg("rcqe=%d\n", rcqe);
 | |
|     pr_dbg("rmax_sge=%d\n", rmax_sge);
 | |
|     pr_dbg("rpages=%d\n", rpages);
 | |
| 
 | |
|     /* Create send ring */
 | |
|     sr->ring_state = (struct pvrdma_ring *)
 | |
|         rdma_pci_dma_map(pci_dev, tbl[0], TARGET_PAGE_SIZE);
 | |
|     if (!sr->ring_state) {
 | |
|         pr_dbg("Failed to map to CQ ring state\n");
 | |
|         goto out_free_sr_mem;
 | |
|     }
 | |
| 
 | |
|     wqe_sz = pow2ceil(sizeof(struct pvrdma_sq_wqe_hdr) +
 | |
|                       sizeof(struct pvrdma_sge) * smax_sge - 1);
 | |
| 
 | |
|     sprintf(ring_name, "qp_sring_%" PRIx64, pdir_dma);
 | |
|     rc = pvrdma_ring_init(sr, ring_name, pci_dev, sr->ring_state,
 | |
|                           scqe, wqe_sz, (dma_addr_t *)&tbl[1], spages);
 | |
|     if (rc) {
 | |
|         goto out_unmap_ring_state;
 | |
|     }
 | |
| 
 | |
|     /* Create recv ring */
 | |
|     rr->ring_state = &sr->ring_state[1];
 | |
|     wqe_sz = pow2ceil(sizeof(struct pvrdma_rq_wqe_hdr) +
 | |
|                       sizeof(struct pvrdma_sge) * rmax_sge - 1);
 | |
|     sprintf(ring_name, "qp_rring_%" PRIx64, pdir_dma);
 | |
|     rc = pvrdma_ring_init(rr, ring_name, pci_dev, rr->ring_state,
 | |
|                           rcqe, wqe_sz, (dma_addr_t *)&tbl[1 + spages], rpages);
 | |
|     if (rc) {
 | |
|         goto out_free_sr;
 | |
|     }
 | |
| 
 | |
|     goto out;
 | |
| 
 | |
| out_free_sr:
 | |
|     pvrdma_ring_free(sr);
 | |
| 
 | |
| out_unmap_ring_state:
 | |
|     rdma_pci_dma_unmap(pci_dev, sr->ring_state, TARGET_PAGE_SIZE);
 | |
| 
 | |
| out_free_sr_mem:
 | |
|     g_free(sr);
 | |
| 
 | |
| out:
 | |
|     rdma_pci_dma_unmap(pci_dev, tbl, TARGET_PAGE_SIZE);
 | |
|     rdma_pci_dma_unmap(pci_dev, dir, TARGET_PAGE_SIZE);
 | |
| 
 | |
|     return rc;
 | |
| }
 | |
| 
 | |
| static int create_qp(PVRDMADev *dev, union pvrdma_cmd_req *req,
 | |
|                      union pvrdma_cmd_resp *rsp)
 | |
| {
 | |
|     struct pvrdma_cmd_create_qp *cmd = &req->create_qp;
 | |
|     struct pvrdma_cmd_create_qp_resp *resp = &rsp->create_qp_resp;
 | |
|     PvrdmaRing *rings = NULL;
 | |
| 
 | |
|     memset(resp, 0, sizeof(*resp));
 | |
|     resp->hdr.response = cmd->hdr.response;
 | |
|     resp->hdr.ack = PVRDMA_CMD_CREATE_QP_RESP;
 | |
| 
 | |
|     pr_dbg("total_chunks=%d\n", cmd->total_chunks);
 | |
|     pr_dbg("send_chunks=%d\n", cmd->send_chunks);
 | |
| 
 | |
|     resp->hdr.err = create_qp_rings(PCI_DEVICE(dev), cmd->pdir_dma, &rings,
 | |
|                                     cmd->max_send_wr, cmd->max_send_sge,
 | |
|                                     cmd->send_chunks, cmd->max_recv_wr,
 | |
|                                     cmd->max_recv_sge, cmd->total_chunks -
 | |
|                                     cmd->send_chunks - 1);
 | |
|     if (resp->hdr.err) {
 | |
|         goto out;
 | |
|     }
 | |
| 
 | |
|     pr_dbg("rings=%p\n", rings);
 | |
| 
 | |
|     resp->hdr.err = rdma_rm_alloc_qp(&dev->rdma_dev_res, cmd->pd_handle,
 | |
|                                      cmd->qp_type, cmd->max_send_wr,
 | |
|                                      cmd->max_send_sge, cmd->send_cq_handle,
 | |
|                                      cmd->max_recv_wr, cmd->max_recv_sge,
 | |
|                                      cmd->recv_cq_handle, rings, &resp->qpn);
 | |
| 
 | |
|     resp->max_send_wr = cmd->max_send_wr;
 | |
|     resp->max_recv_wr = cmd->max_recv_wr;
 | |
|     resp->max_send_sge = cmd->max_send_sge;
 | |
|     resp->max_recv_sge = cmd->max_recv_sge;
 | |
|     resp->max_inline_data = cmd->max_inline_data;
 | |
| 
 | |
| out:
 | |
|     pr_dbg("ret=%d\n", resp->hdr.err);
 | |
|     return resp->hdr.err;
 | |
| }
 | |
| 
 | |
| static int modify_qp(PVRDMADev *dev, union pvrdma_cmd_req *req,
 | |
|                      union pvrdma_cmd_resp *rsp)
 | |
| {
 | |
|     struct pvrdma_cmd_modify_qp *cmd = &req->modify_qp;
 | |
| 
 | |
|     pr_dbg("qp_handle=%d\n", cmd->qp_handle);
 | |
| 
 | |
|     memset(rsp, 0, sizeof(*rsp));
 | |
|     rsp->hdr.response = cmd->hdr.response;
 | |
|     rsp->hdr.ack = PVRDMA_CMD_MODIFY_QP_RESP;
 | |
| 
 | |
|     rsp->hdr.err = rdma_rm_modify_qp(&dev->rdma_dev_res, &dev->backend_dev,
 | |
|                                  cmd->qp_handle, cmd->attr_mask,
 | |
|                                  (union ibv_gid *)&cmd->attrs.ah_attr.grh.dgid,
 | |
|                                  cmd->attrs.dest_qp_num,
 | |
|                                  (enum ibv_qp_state)cmd->attrs.qp_state,
 | |
|                                  cmd->attrs.qkey, cmd->attrs.rq_psn,
 | |
|                                  cmd->attrs.sq_psn);
 | |
| 
 | |
|     pr_dbg("ret=%d\n", rsp->hdr.err);
 | |
|     return rsp->hdr.err;
 | |
| }
 | |
| 
 | |
| static int query_qp(PVRDMADev *dev, union pvrdma_cmd_req *req,
 | |
|                      union pvrdma_cmd_resp *rsp)
 | |
| {
 | |
|     struct pvrdma_cmd_query_qp *cmd = &req->query_qp;
 | |
|     struct pvrdma_cmd_query_qp_resp *resp = &rsp->query_qp_resp;
 | |
|     struct ibv_qp_init_attr init_attr;
 | |
| 
 | |
|     pr_dbg("qp_handle=%d\n", cmd->qp_handle);
 | |
| 
 | |
|     memset(rsp, 0, sizeof(*rsp));
 | |
|     rsp->hdr.response = cmd->hdr.response;
 | |
|     rsp->hdr.ack = PVRDMA_CMD_QUERY_QP_RESP;
 | |
| 
 | |
|     rsp->hdr.err = rdma_rm_query_qp(&dev->rdma_dev_res, &dev->backend_dev,
 | |
|                                     cmd->qp_handle,
 | |
|                                     (struct ibv_qp_attr *)&resp->attrs, -1,
 | |
|                                     &init_attr);
 | |
| 
 | |
|     pr_dbg("ret=%d\n", rsp->hdr.err);
 | |
|     return rsp->hdr.err;
 | |
| }
 | |
| 
 | |
| static int destroy_qp(PVRDMADev *dev, union pvrdma_cmd_req *req,
 | |
|                       union pvrdma_cmd_resp *rsp)
 | |
| {
 | |
|     struct pvrdma_cmd_destroy_qp *cmd = &req->destroy_qp;
 | |
|     RdmaRmQP *qp;
 | |
|     PvrdmaRing *ring;
 | |
| 
 | |
|     qp = rdma_rm_get_qp(&dev->rdma_dev_res, cmd->qp_handle);
 | |
|     if (!qp) {
 | |
|         pr_dbg("Invalid QP handle\n");
 | |
|         return -EINVAL;
 | |
|     }
 | |
| 
 | |
|     rdma_rm_dealloc_qp(&dev->rdma_dev_res, cmd->qp_handle);
 | |
| 
 | |
|     ring = (PvrdmaRing *)qp->opaque;
 | |
|     pr_dbg("sring=%p\n", &ring[0]);
 | |
|     pvrdma_ring_free(&ring[0]);
 | |
|     pr_dbg("rring=%p\n", &ring[1]);
 | |
|     pvrdma_ring_free(&ring[1]);
 | |
| 
 | |
|     rdma_pci_dma_unmap(PCI_DEVICE(dev), ring->ring_state, TARGET_PAGE_SIZE);
 | |
|     g_free(ring);
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static int create_bind(PVRDMADev *dev, union pvrdma_cmd_req *req,
 | |
|                        union pvrdma_cmd_resp *rsp)
 | |
| {
 | |
|     struct pvrdma_cmd_create_bind *cmd = &req->create_bind;
 | |
| #ifdef PVRDMA_DEBUG
 | |
|     __be64 *subnet = (__be64 *)&cmd->new_gid[0];
 | |
|     __be64 *if_id = (__be64 *)&cmd->new_gid[8];
 | |
| #endif
 | |
| 
 | |
|     pr_dbg("index=%d\n", cmd->index);
 | |
| 
 | |
|     if (cmd->index >= MAX_PORT_GIDS) {
 | |
|         return -EINVAL;
 | |
|     }
 | |
| 
 | |
|     pr_dbg("gid[%d]=0x%llx,0x%llx\n", cmd->index,
 | |
|            (long long unsigned int)be64_to_cpu(*subnet),
 | |
|            (long long unsigned int)be64_to_cpu(*if_id));
 | |
| 
 | |
|     /* Driver forces to one port only */
 | |
|     memcpy(dev->rdma_dev_res.ports[0].gid_tbl[cmd->index].raw, &cmd->new_gid,
 | |
|            sizeof(cmd->new_gid));
 | |
| 
 | |
|     /* TODO: Since drivers stores node_guid at load_dsr phase then this
 | |
|      * assignment is not relevant, i need to figure out a way how to
 | |
|      * retrieve MAC of our netdev */
 | |
|     dev->node_guid = dev->rdma_dev_res.ports[0].gid_tbl[0].global.interface_id;
 | |
|     pr_dbg("dev->node_guid=0x%llx\n",
 | |
|            (long long unsigned int)be64_to_cpu(dev->node_guid));
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static int destroy_bind(PVRDMADev *dev, union pvrdma_cmd_req *req,
 | |
|                         union pvrdma_cmd_resp *rsp)
 | |
| {
 | |
|     struct pvrdma_cmd_destroy_bind *cmd = &req->destroy_bind;
 | |
| 
 | |
|     pr_dbg("index=%d\n", cmd->index);
 | |
| 
 | |
|     if (cmd->index >= MAX_PORT_GIDS) {
 | |
|         return -EINVAL;
 | |
|     }
 | |
| 
 | |
|     memset(dev->rdma_dev_res.ports[0].gid_tbl[cmd->index].raw, 0,
 | |
|            sizeof(dev->rdma_dev_res.ports[0].gid_tbl[cmd->index].raw));
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static int create_uc(PVRDMADev *dev, union pvrdma_cmd_req *req,
 | |
|                      union pvrdma_cmd_resp *rsp)
 | |
| {
 | |
|     struct pvrdma_cmd_create_uc *cmd = &req->create_uc;
 | |
|     struct pvrdma_cmd_create_uc_resp *resp = &rsp->create_uc_resp;
 | |
| 
 | |
|     pr_dbg("pfn=%d\n", cmd->pfn);
 | |
| 
 | |
|     memset(resp, 0, sizeof(*resp));
 | |
|     resp->hdr.response = cmd->hdr.response;
 | |
|     resp->hdr.ack = PVRDMA_CMD_CREATE_UC_RESP;
 | |
|     resp->hdr.err = rdma_rm_alloc_uc(&dev->rdma_dev_res, cmd->pfn,
 | |
|                                      &resp->ctx_handle);
 | |
| 
 | |
|     pr_dbg("ret=%d\n", resp->hdr.err);
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static int destroy_uc(PVRDMADev *dev, union pvrdma_cmd_req *req,
 | |
|                       union pvrdma_cmd_resp *rsp)
 | |
| {
 | |
|     struct pvrdma_cmd_destroy_uc *cmd = &req->destroy_uc;
 | |
| 
 | |
|     pr_dbg("ctx_handle=%d\n", cmd->ctx_handle);
 | |
| 
 | |
|     rdma_rm_dealloc_uc(&dev->rdma_dev_res, cmd->ctx_handle);
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| struct cmd_handler {
 | |
|     uint32_t cmd;
 | |
|     int (*exec)(PVRDMADev *dev, union pvrdma_cmd_req *req,
 | |
|             union pvrdma_cmd_resp *rsp);
 | |
| };
 | |
| 
 | |
| static struct cmd_handler cmd_handlers[] = {
 | |
|     {PVRDMA_CMD_QUERY_PORT, query_port},
 | |
|     {PVRDMA_CMD_QUERY_PKEY, query_pkey},
 | |
|     {PVRDMA_CMD_CREATE_PD, create_pd},
 | |
|     {PVRDMA_CMD_DESTROY_PD, destroy_pd},
 | |
|     {PVRDMA_CMD_CREATE_MR, create_mr},
 | |
|     {PVRDMA_CMD_DESTROY_MR, destroy_mr},
 | |
|     {PVRDMA_CMD_CREATE_CQ, create_cq},
 | |
|     {PVRDMA_CMD_RESIZE_CQ, NULL},
 | |
|     {PVRDMA_CMD_DESTROY_CQ, destroy_cq},
 | |
|     {PVRDMA_CMD_CREATE_QP, create_qp},
 | |
|     {PVRDMA_CMD_MODIFY_QP, modify_qp},
 | |
|     {PVRDMA_CMD_QUERY_QP, query_qp},
 | |
|     {PVRDMA_CMD_DESTROY_QP, destroy_qp},
 | |
|     {PVRDMA_CMD_CREATE_UC, create_uc},
 | |
|     {PVRDMA_CMD_DESTROY_UC, destroy_uc},
 | |
|     {PVRDMA_CMD_CREATE_BIND, create_bind},
 | |
|     {PVRDMA_CMD_DESTROY_BIND, destroy_bind},
 | |
| };
 | |
| 
 | |
| int execute_command(PVRDMADev *dev)
 | |
| {
 | |
|     int err = 0xFFFF;
 | |
|     DSRInfo *dsr_info;
 | |
| 
 | |
|     dsr_info = &dev->dsr_info;
 | |
| 
 | |
|     pr_dbg("cmd=%d\n", dsr_info->req->hdr.cmd);
 | |
|     if (dsr_info->req->hdr.cmd >= sizeof(cmd_handlers) /
 | |
|                       sizeof(struct cmd_handler)) {
 | |
|         pr_dbg("Unsupported command\n");
 | |
|         goto out;
 | |
|     }
 | |
| 
 | |
|     if (!cmd_handlers[dsr_info->req->hdr.cmd].exec) {
 | |
|         pr_dbg("Unsupported command (not implemented yet)\n");
 | |
|         goto out;
 | |
|     }
 | |
| 
 | |
|     err = cmd_handlers[dsr_info->req->hdr.cmd].exec(dev, dsr_info->req,
 | |
|                             dsr_info->rsp);
 | |
| out:
 | |
|     set_reg_val(dev, PVRDMA_REG_ERR, err);
 | |
|     post_interrupt(dev, INTR_VEC_CMD_RING);
 | |
| 
 | |
|     return (err == 0) ? 0 : -EINVAL;
 | |
| }
 |