gve: fix dma sync bug where not all pages synced
authorAdi Suresh <adisuresh@google.com>
Tue, 19 Nov 2019 16:02:47 +0000 (08:02 -0800)
committerDavid S. Miller <davem@davemloft.net>
Tue, 19 Nov 2019 20:58:18 +0000 (12:58 -0800)
The previous commit had a bug where the last page in the memory range
could not be synced. This change fixes the behavior so that all the
required pages are synced.

Fixes: 9cfeeb576d49 ("gve: Fixes DMA synchronization")
Signed-off-by: Adi Suresh <adisuresh@google.com>
Reviewed-by: Catherine Sullivan <csully@google.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
drivers/net/ethernet/google/gve/gve_tx.c

index 0a9a7ee2a866838cce439172af87edc6ef0106e7..f4889431f9b7049f0046338a3bed0884ade84782 100644 (file)
@@ -393,12 +393,13 @@ static void gve_tx_fill_seg_desc(union gve_tx_desc *seg_desc,
 static void gve_dma_sync_for_device(struct device *dev, dma_addr_t *page_buses,
                                    u64 iov_offset, u64 iov_len)
 {
+       u64 last_page = (iov_offset + iov_len - 1) / PAGE_SIZE;
+       u64 first_page = iov_offset / PAGE_SIZE;
        dma_addr_t dma;
-       u64 addr;
+       u64 page;
 
-       for (addr = iov_offset; addr < iov_offset + iov_len;
-            addr += PAGE_SIZE) {
-               dma = page_buses[addr / PAGE_SIZE];
+       for (page = first_page; page <= last_page; page++) {
+               dma = page_buses[page];
                dma_sync_single_for_device(dev, dma, PAGE_SIZE, DMA_TO_DEVICE);
        }
 }