[dpdk-dev] [PATCH 08/21] vhost: iotlb: add pending miss request list and helpers
Tiwei Bie
tiwei.bie at intel.com
Tue Sep 5 09:11:06 CEST 2017
On Thu, Aug 31, 2017 at 11:50:10AM +0200, Maxime Coquelin wrote:
> In order to be able to handle other ports or queues while waiting
> for an IOTLB miss reply, a pending list is created so that waiter
> can return and restart later on with sending again a miss request.
>
> Signed-off-by: Maxime Coquelin <maxime.coquelin at redhat.com>
> ---
> lib/librte_vhost/iotlb.c | 88 ++++++++++++++++++++++++++++++++++++++++++++++--
> lib/librte_vhost/iotlb.h | 4 +++
> lib/librte_vhost/vhost.h | 1 +
> 3 files changed, 91 insertions(+), 2 deletions(-)
>
> diff --git a/lib/librte_vhost/iotlb.c b/lib/librte_vhost/iotlb.c
> index 1b739dae5..d014bfe98 100644
> --- a/lib/librte_vhost/iotlb.c
> +++ b/lib/librte_vhost/iotlb.c
> @@ -49,7 +49,86 @@ struct vhost_iotlb_entry {
> uint8_t perm;
> };
>
> -#define IOTLB_CACHE_SIZE 1024
> +#define IOTLB_CACHE_SIZE 2048
> +
> +static void vhost_user_iotlb_pending_remove_all(struct vhost_virtqueue *vq)
> +{
> + struct vhost_iotlb_entry *node, *temp_node;
> +
> + rte_rwlock_write_lock(&vq->iotlb_lock);
> +
> + TAILQ_FOREACH_SAFE(node, &vq->iotlb_pending_list, next, temp_node) {
> + TAILQ_REMOVE(&vq->iotlb_pending_list, node, next);
> + rte_mempool_put(vq->iotlb_pool, node);
> + }
> +
> + rte_rwlock_write_unlock(&vq->iotlb_lock);
> +}
> +
> +int vhost_user_iotlb_pending_miss(struct vhost_virtqueue *vq, uint64_t iova,
> + uint8_t perm)
> +{
> + struct vhost_iotlb_entry *node;
> + int found = 0;
> +
The return value of this function is boolean. So it's better
to return bool instead of int.
> + rte_rwlock_read_lock(&vq->iotlb_lock);
> +
> + TAILQ_FOREACH(node, &vq->iotlb_pending_list, next) {
> + if ((node->iova == iova) && (node->perm == perm)) {
> + found = 1;
> + break;
> + }
> + }
> +
> + rte_rwlock_read_unlock(&vq->iotlb_lock);
> +
> + return found;
> +}
> +
> +void vhost_user_iotlb_pending_insert(struct vhost_virtqueue *vq,
> + uint64_t iova, uint8_t perm)
> +{
> + struct vhost_iotlb_entry *node;
> + int ret;
> +
> + ret = rte_mempool_get(vq->iotlb_pool, (void **)&node);
> + if (ret) {
> + RTE_LOG(ERR, VHOST_CONFIG, "IOTLB pool empty, invalidate cache\n");
I think The log level should be INFO or the likes, not ERR.
> + vhost_user_iotlb_pending_remove_all(vq);
> + ret = rte_mempool_get(vq->iotlb_pool, (void **)&node);
> + if (ret) {
> + RTE_LOG(ERR, VHOST_CONFIG, "IOTLB pool still empty, failure\n");
> + return;
> + }
> + }
> +
> + node->iova = iova;
> + node->perm = perm;
> +
> + rte_rwlock_write_lock(&vq->iotlb_lock);
> +
> + TAILQ_INSERT_TAIL(&vq->iotlb_pending_list, node, next);
> +
> + rte_rwlock_write_unlock(&vq->iotlb_lock);
> +}
> +
> +static void vhost_user_iotlb_pending_remove(struct vhost_virtqueue *vq,
> + uint64_t iova, uint64_t size, uint8_t perm)
> +{
> + struct vhost_iotlb_entry *node, *temp_node;
> +
> + /* .iotlb_lock already locked by the caller */
> + TAILQ_FOREACH_SAFE(node, &vq->iotlb_pending_list, next, temp_node) {
> + if (node->iova < iova)
> + continue;
> + if (node->iova >= iova + size)
> + continue;
> + if ((node->perm & perm) != node->perm)
> + continue;
> + TAILQ_REMOVE(&vq->iotlb_pending_list, node, next);
> + rte_mempool_put(vq->iotlb_pool, node);
> + }
> +}
>
> static void vhost_user_iotlb_cache_remove_all(struct vhost_virtqueue *vq)
> {
> @@ -106,7 +185,10 @@ void vhost_user_iotlb_cache_insert(struct vhost_virtqueue *vq, uint64_t iova,
> TAILQ_INSERT_TAIL(&vq->iotlb_list, new_node, next);
>
> unlock:
> + vhost_user_iotlb_pending_remove(vq, iova, size, perm);
> +
> rte_rwlock_write_unlock(&vq->iotlb_lock);
> +
This empty line should be removed.
Best regards,
Tiwei Bie
> }
>
> void vhost_user_iotlb_cache_remove(struct vhost_virtqueue *vq,
More information about the dev
mailing list