From mboxrd@z Thu Jan 1 00:00:00 1970 From: Yuanhan Liu Subject: Re: [PATCH v3 2/2] vhost: Add VHOST PMD Date: Mon, 9 Nov 2015 14:21:42 +0800 Message-ID: <20151109062142.GN2326@yliu-dev.sh.intel.com> References: <1446436737-25606-2-git-send-email-mukawa@igel.co.jp> <1447046221-20811-1-git-send-email-mukawa@igel.co.jp> <1447046221-20811-3-git-send-email-mukawa@igel.co.jp> Mime-Version: 1.0 Content-Type: text/plain; charset=us-ascii Cc: dev@dpdk.org, ann.zhuangyanying@huawei.com To: Tetsuya Mukawa Return-path: Received: from mga11.intel.com (mga11.intel.com [192.55.52.93]) by dpdk.org (Postfix) with ESMTP id 8E1A03979 for ; Mon, 9 Nov 2015 07:18:12 +0100 (CET) Content-Disposition: inline In-Reply-To: <1447046221-20811-3-git-send-email-mukawa@igel.co.jp> List-Id: patches and discussions about DPDK List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" Hi Tetsuya, Here I just got some minor nits after a very rough glimpse. On Mon, Nov 09, 2015 at 02:17:01PM +0900, Tetsuya Mukawa wrote: ... > +static uint16_t > +eth_vhost_rx(void *q, struct rte_mbuf **bufs, uint16_t nb_bufs) > +{ > + struct vhost_queue *r = q; > + uint16_t nb_rx = 0; > + > + if (unlikely(rte_atomic32_read(&r->allow_queuing) == 0)) > + return 0; > + > + rte_atomic32_set(&r->while_queuing, 1); > + > + if (unlikely(rte_atomic32_read(&r->allow_queuing) == 0)) > + goto out; > + > + /* Dequeue packets from guest TX queue */ > + nb_rx = (uint16_t)rte_vhost_dequeue_burst(r->device, > + r->virtqueue_id, r->mb_pool, bufs, nb_bufs); Unnecessary cast, as rte_vhost_enqueue_burst is defined with uint16_t return type. > + > + r->rx_pkts += nb_rx; > + > +out: > + rte_atomic32_set(&r->while_queuing, 0); > + > + return nb_rx; > +} > + > +static uint16_t > +eth_vhost_tx(void *q, struct rte_mbuf **bufs, uint16_t nb_bufs) > +{ > + struct vhost_queue *r = q; > + uint16_t i, nb_tx = 0; > + > + if (unlikely(rte_atomic32_read(&r->allow_queuing) == 0)) > + return 0; > + > + rte_atomic32_set(&r->while_queuing, 1); > + > + if (unlikely(rte_atomic32_read(&r->allow_queuing) == 0)) > + goto out; > + > + /* Enqueue packets to guest RX queue */ > + nb_tx = (uint16_t)rte_vhost_enqueue_burst(r->device, > + r->virtqueue_id, bufs, nb_bufs); Ditto. > + > + r->tx_pkts += nb_tx; > + r->err_pkts += nb_bufs - nb_tx; > + > + for (i = 0; likely(i < nb_tx); i++) > + rte_pktmbuf_free(bufs[i]); > + > +out: > + rte_atomic32_set(&r->while_queuing, 0); > + > + return nb_tx; > +} > + > +static int > +eth_dev_configure(struct rte_eth_dev *dev __rte_unused) { return 0; } I personally would not prefer to saving few lines of code to sacrifice the readability. > + > +static int > +eth_rx_queue_setup(struct rte_eth_dev *dev, uint16_t rx_queue_id, > + uint16_t nb_rx_desc __rte_unused, > + unsigned int socket_id, > + const struct rte_eth_rxconf *rx_conf __rte_unused, > + struct rte_mempool *mb_pool) > +{ > + struct pmd_internal *internal = dev->data->dev_private; > + struct vhost_queue *vq; > + > + if (internal->rx_vhost_queues[rx_queue_id] != NULL) > + rte_free(internal->rx_vhost_queues[rx_queue_id]); Such NULL check is unnecessary; rte_free will handle it. > + > + vq = rte_zmalloc_socket(NULL, sizeof(struct vhost_queue), > + RTE_CACHE_LINE_SIZE, socket_id); > + if (vq == NULL) { > + RTE_LOG(ERR, PMD, "Failed to allocate memory for rx queue\n"); > + return -ENOMEM; > + } > + > + vq->mb_pool = mb_pool; > + vq->virtqueue_id = rx_queue_id * VIRTIO_QNUM + VIRTIO_TXQ; > + internal->rx_vhost_queues[rx_queue_id] = vq; > + dev->data->rx_queues[rx_queue_id] = vq; > + return 0; > +} > + > +static int > +eth_tx_queue_setup(struct rte_eth_dev *dev, uint16_t tx_queue_id, > + uint16_t nb_tx_desc __rte_unused, > + unsigned int socket_id, > + const struct rte_eth_txconf *tx_conf __rte_unused) > +{ > + struct pmd_internal *internal = dev->data->dev_private; > + struct vhost_queue *vq; > + > + if (internal->tx_vhost_queues[tx_queue_id] != NULL) > + rte_free(internal->tx_vhost_queues[tx_queue_id]); Ditto. > + > + vq = rte_zmalloc_socket(NULL, sizeof(struct vhost_queue), > + RTE_CACHE_LINE_SIZE, socket_id); > + if (vq == NULL) { > + RTE_LOG(ERR, PMD, "Failed to allocate memory for tx queue\n"); > + return -ENOMEM; > + } > + > + vq->virtqueue_id = tx_queue_id * VIRTIO_QNUM + VIRTIO_RXQ; > + internal->tx_vhost_queues[tx_queue_id] = vq; > + dev->data->tx_queues[tx_queue_id] = vq; > + return 0; > +} > + > + > +static void > +eth_dev_info(struct rte_eth_dev *dev, > + struct rte_eth_dev_info *dev_info) > +{ > + struct pmd_internal *internal = dev->data->dev_private; > + > + dev_info->driver_name = drivername; > + dev_info->max_mac_addrs = 1; > + dev_info->max_rx_pktlen = (uint32_t)-1; > + dev_info->max_rx_queues = (uint16_t)internal->nb_rx_queues; > + dev_info->max_tx_queues = (uint16_t)internal->nb_tx_queues; > + dev_info->min_rx_bufsize = 0; > +} > + > +static void > +eth_stats_get(struct rte_eth_dev *dev, struct rte_eth_stats *igb_stats) > +{ > + unsigned i; > + unsigned long rx_total = 0, tx_total = 0, tx_err_total = 0; > + const struct pmd_internal *internal = dev->data->dev_private; > + > + for (i = 0; i < RTE_ETHDEV_QUEUE_STAT_CNTRS && > + i < internal->nb_rx_queues; i++) { > + if (internal->rx_vhost_queues[i] == NULL) > + continue; > + igb_stats->q_ipackets[i] = internal->rx_vhost_queues[i]->rx_pkts; > + rx_total += igb_stats->q_ipackets[i]; > + } > + > + for (i = 0; i < RTE_ETHDEV_QUEUE_STAT_CNTRS && > + i < internal->nb_tx_queues; i++) { > + if (internal->tx_vhost_queues[i] == NULL) > + continue; > + igb_stats->q_opackets[i] = internal->tx_vhost_queues[i]->tx_pkts; > + igb_stats->q_errors[i] = internal->tx_vhost_queues[i]->err_pkts; > + tx_total += igb_stats->q_opackets[i]; > + tx_err_total += igb_stats->q_errors[i]; > + } > + > + igb_stats->ipackets = rx_total; > + igb_stats->opackets = tx_total; > + igb_stats->oerrors = tx_err_total; > +} > + > +static void > +eth_stats_reset(struct rte_eth_dev *dev) > +{ > + unsigned i; > + struct pmd_internal *internal = dev->data->dev_private; > + > + for (i = 0; i < internal->nb_rx_queues; i++) { > + if (internal->rx_vhost_queues[i] == NULL) > + continue; > + internal->rx_vhost_queues[i]->rx_pkts = 0; > + } > + for (i = 0; i < internal->nb_tx_queues; i++) { > + if (internal->tx_vhost_queues[i] == NULL) > + continue; > + internal->tx_vhost_queues[i]->tx_pkts = 0; > + internal->tx_vhost_queues[i]->err_pkts = 0; > + } > +} > + > +static void > +eth_queue_release(void *q __rte_unused) { ; } > +static int > +eth_link_update(struct rte_eth_dev *dev __rte_unused, > + int wait_to_complete __rte_unused) { return 0; } Ditto. > + > +static const struct eth_dev_ops ops = { > + .dev_start = eth_dev_start, > + .dev_stop = eth_dev_stop, > + .dev_configure = eth_dev_configure, > + .dev_infos_get = eth_dev_info, > + .rx_queue_setup = eth_rx_queue_setup, > + .tx_queue_setup = eth_tx_queue_setup, > + .rx_queue_release = eth_queue_release, > + .tx_queue_release = eth_queue_release, > + .link_update = eth_link_update, > + .stats_get = eth_stats_get, > + .stats_reset = eth_stats_reset, > +}; > + > +static int > +eth_dev_vhost_create(const char *name, int index, > + char *iface_name, > + int16_t queues, > + const unsigned numa_node) > +{ > + struct rte_eth_dev_data *data = NULL; > + struct pmd_internal *internal = NULL; > + struct rte_eth_dev *eth_dev = NULL; > + struct ether_addr *eth_addr = NULL; > + > + RTE_LOG(INFO, PMD, "Creating VHOST-USER backend on numa socket %u\n", > + numa_node); > + > + /* now do all data allocation - for eth_dev structure, dummy pci driver > + * and internal (private) data > + */ > + data = rte_zmalloc_socket(name, sizeof(*data), 0, numa_node); > + if (data == NULL) > + goto error; > + > + internal = rte_zmalloc_socket(name, sizeof(*internal), 0, numa_node); > + if (internal == NULL) > + goto error; > + > + eth_addr = rte_zmalloc_socket(name, sizeof(*eth_addr), 0, numa_node); > + if (eth_addr == NULL) > + goto error; > + *eth_addr = base_eth_addr; > + eth_addr->addr_bytes[5] = index; > + > + /* reserve an ethdev entry */ > + eth_dev = rte_eth_dev_allocate(name, RTE_ETH_DEV_VIRTUAL); > + if (eth_dev == NULL) > + goto error; > + > + /* now put it all together > + * - store queue data in internal, > + * - store numa_node info in ethdev data > + * - point eth_dev_data to internals > + * - and point eth_dev structure to new eth_dev_data structure > + */ > + internal->nb_rx_queues = queues; > + internal->nb_tx_queues = queues; > + internal->dev_name = strdup(name); > + if (internal->dev_name == NULL) > + goto error; > + internal->iface_name = strdup(iface_name); > + if (internal->iface_name == NULL) > + goto error; If allocation failed here, you will find that internal->dev_name is not freed. > + > + pthread_mutex_lock(&internal_list_lock); > + TAILQ_INSERT_TAIL(&internals_list, internal, next); > + pthread_mutex_unlock(&internal_list_lock); > + > + data->dev_private = internal; > + data->port_id = eth_dev->data->port_id; > + memmove(data->name, eth_dev->data->name, sizeof(data->name)); > + data->nb_rx_queues = queues; > + data->nb_tx_queues = queues; > + data->dev_link = pmd_link; > + data->mac_addrs = eth_addr; > + > + /* We'll replace the 'data' originally allocated by eth_dev. So the > + * vhost PMD resources won't be shared between multi processes. > + */ > + eth_dev->data = data; > + eth_dev->dev_ops = &ops; > + eth_dev->driver = NULL; > + eth_dev->data->dev_flags = RTE_ETH_DEV_DETACHABLE; > + eth_dev->data->kdrv = RTE_KDRV_NONE; > + eth_dev->data->drv_name = internal->dev_name; > + eth_dev->data->numa_node = numa_node; > + > + /* finally assign rx and tx ops */ > + eth_dev->rx_pkt_burst = eth_vhost_rx; > + eth_dev->tx_pkt_burst = eth_vhost_tx; > + > + return data->port_id; > + > +error: > + rte_free(data); > + rte_free(internal); > + rte_free(eth_addr); > + > + return -1; > +} ... ... > + > + if ((internal) && (internal->dev_name)) > + free(internal->dev_name); > + if ((internal) && (internal->iface_name)) > + free(internal->iface_name); > + > + rte_free(eth_dev->data->mac_addrs); > + rte_free(eth_dev->data); > + > + for (i = 0; i < internal->nb_rx_queues; i++) { > + if (internal->rx_vhost_queues[i] != NULL) > + rte_free(internal->rx_vhost_queues[i]); > + } > + for (i = 0; i < internal->nb_tx_queues; i++) { > + if (internal->tx_vhost_queues[i] != NULL) > + rte_free(internal->tx_vhost_queues[i]); Ditto. (Hopefully I could have a detailed review later, say next week). --yliu