iommu/iova: Add flush counters to Flush-Queue implementation
authorJoerg Roedel <jroedel@suse.de>
Thu, 10 Aug 2017 14:14:59 +0000 (16:14 +0200)
committerJoerg Roedel <jroedel@suse.de>
Tue, 15 Aug 2017 16:23:51 +0000 (18:23 +0200)
There are two counters:

* fq_flush_start_cnt  - Increased when a TLB flush
                        is started.

* fq_flush_finish_cnt - Increased when a TLB flush
is finished.

The fq_flush_start_cnt is assigned to every Flush-Queue
entry on its creation. When freeing entries from the
Flush-Queue, the value in the entry is compared to the
fq_flush_finish_cnt. The entry can only be freed when its
value is less than the value of fq_flush_finish_cnt.

The reason for these counters it to take advantage of IOMMU
TLB flushes that happened on other CPUs. These already
flushed the TLB for Flush-Queue entries on other CPUs so
that they can already be freed without flushing the TLB
again.

This makes it less likely that the Flush-Queue is full and
saves IOMMU TLB flushes.

Signed-off-by: Joerg Roedel <jroedel@suse.de>
drivers/iommu/iova.c
include/linux/iova.h

index e5c9a7ae6088a0cb9ceaa25fd8c86be68674aa45..47b144e417ad6d1d5192e7349d5eaa32e4df2728 100644 (file)
@@ -75,6 +75,9 @@ int init_iova_flush_queue(struct iova_domain *iovad,
 {
        int cpu;
 
+       atomic64_set(&iovad->fq_flush_start_cnt,  0);
+       atomic64_set(&iovad->fq_flush_finish_cnt, 0);
+
        iovad->fq = alloc_percpu(struct iova_fq);
        if (!iovad->fq)
                return -ENOMEM;
@@ -482,20 +485,30 @@ static inline unsigned fq_ring_add(struct iova_fq *fq)
 
 static void fq_ring_free(struct iova_domain *iovad, struct iova_fq *fq)
 {
+       u64 counter = atomic64_read(&iovad->fq_flush_finish_cnt);
        unsigned idx;
 
        fq_ring_for_each(idx, fq) {
 
+               if (fq->entries[idx].counter >= counter)
+                       break;
+
                if (iovad->entry_dtor)
                        iovad->entry_dtor(fq->entries[idx].data);
 
                free_iova_fast(iovad,
                               fq->entries[idx].iova_pfn,
                               fq->entries[idx].pages);
+
+               fq->head = (fq->head + 1) % IOVA_FQ_SIZE;
        }
+}
 
-       fq->head = 0;
-       fq->tail = 0;
+static void iova_domain_flush(struct iova_domain *iovad)
+{
+       atomic64_inc(&iovad->fq_flush_start_cnt);
+       iovad->flush_cb(iovad);
+       atomic64_inc(&iovad->fq_flush_finish_cnt);
 }
 
 static void fq_destroy_all_entries(struct iova_domain *iovad)
@@ -526,8 +539,15 @@ void queue_iova(struct iova_domain *iovad,
        struct iova_fq *fq = get_cpu_ptr(iovad->fq);
        unsigned idx;
 
+       /*
+        * First remove all entries from the flush queue that have already been
+        * flushed out on another CPU. This makes the fq_full() check below less
+        * likely to be true.
+        */
+       fq_ring_free(iovad, fq);
+
        if (fq_full(fq)) {
-               iovad->flush_cb(iovad);
+               iova_domain_flush(iovad);
                fq_ring_free(iovad, fq);
        }
 
@@ -536,6 +556,7 @@ void queue_iova(struct iova_domain *iovad,
        fq->entries[idx].iova_pfn = pfn;
        fq->entries[idx].pages    = pages;
        fq->entries[idx].data     = data;
+       fq->entries[idx].counter  = atomic64_read(&iovad->fq_flush_start_cnt);
 
        put_cpu_ptr(iovad->fq);
 }
index 1ae85248ec503c936b5f692824228dad238cbc5a..985b8008999ec777f502aa1124d047fb89391601 100644 (file)
@@ -14,6 +14,7 @@
 #include <linux/types.h>
 #include <linux/kernel.h>
 #include <linux/rbtree.h>
+#include <linux/atomic.h>
 #include <linux/dma-mapping.h>
 
 /* iova structure */
@@ -52,6 +53,7 @@ struct iova_fq_entry {
        unsigned long iova_pfn;
        unsigned long pages;
        unsigned long data;
+       u64 counter; /* Flush counter when this entrie was added */
 };
 
 /* Per-CPU Flush Queue structure */
@@ -77,6 +79,12 @@ struct iova_domain {
                                           iova entry */
 
        struct iova_fq __percpu *fq;    /* Flush Queue */
+
+       atomic64_t      fq_flush_start_cnt;     /* Number of TLB flushes that
+                                                  have been started */
+
+       atomic64_t      fq_flush_finish_cnt;    /* Number of TLB flushes that
+                                                  have been finished */
 };
 
 static inline unsigned long iova_size(struct iova *iova)