iommu/arm-smmu: Track context bank state
authorRobin Murphy <robin.murphy@arm.com>
Tue, 8 Aug 2017 13:56:14 +0000 (14:56 +0100)
committerWill Deacon <will.deacon@arm.com>
Wed, 16 Aug 2017 16:27:28 +0000 (17:27 +0100)
Echoing what we do for Stream Map Entries, maintain a software shadow
state for context bank configuration. With this in place, we are mere
moments away from blissfully easy suspend/resume support.

Reviewed-by: Sricharan R <sricharan@codeaurora.org>
Signed-off-by: Robin Murphy <robin.murphy@arm.com>
[will: fix sparse warning by only clearing .cfg during domain destruction]
Signed-off-by: Will Deacon <will.deacon@arm.com>
drivers/iommu/arm-smmu.c

index b0b126ea9d857f69e13d0980f1603b8c81e5ca1e..0a5fa126f3790f6729316434997cbf338807a49f 100644 (file)
@@ -338,6 +338,13 @@ struct arm_smmu_smr {
        bool                            valid;
 };
 
+struct arm_smmu_cb {
+       u64                             ttbr[2];
+       u32                             tcr[2];
+       u32                             mair[2];
+       struct arm_smmu_cfg             *cfg;
+};
+
 struct arm_smmu_master_cfg {
        struct arm_smmu_device          *smmu;
        s16                             smendx[];
@@ -380,6 +387,7 @@ struct arm_smmu_device {
        u32                             num_context_banks;
        u32                             num_s2_context_banks;
        DECLARE_BITMAP(context_map, ARM_SMMU_MAX_CBS);
+       struct arm_smmu_cb              *cbs;
        atomic_t                        irptndx;
 
        u32                             num_mapping_groups;
@@ -776,17 +784,74 @@ static irqreturn_t arm_smmu_global_fault(int irq, void *dev)
 static void arm_smmu_init_context_bank(struct arm_smmu_domain *smmu_domain,
                                       struct io_pgtable_cfg *pgtbl_cfg)
 {
-       u32 reg, reg2;
-       u64 reg64;
-       bool stage1;
        struct arm_smmu_cfg *cfg = &smmu_domain->cfg;
-       struct arm_smmu_device *smmu = smmu_domain->smmu;
+       struct arm_smmu_cb *cb = &smmu_domain->smmu->cbs[cfg->cbndx];
+       bool stage1 = cfg->cbar != CBAR_TYPE_S2_TRANS;
+
+       cb->cfg = cfg;
+
+       /* TTBCR */
+       if (stage1) {
+               if (cfg->fmt == ARM_SMMU_CTX_FMT_AARCH32_S) {
+                       cb->tcr[0] = pgtbl_cfg->arm_v7s_cfg.tcr;
+               } else {
+                       cb->tcr[0] = pgtbl_cfg->arm_lpae_s1_cfg.tcr;
+                       cb->tcr[1] = pgtbl_cfg->arm_lpae_s1_cfg.tcr >> 32;
+                       cb->tcr[1] |= TTBCR2_SEP_UPSTREAM;
+                       if (cfg->fmt == ARM_SMMU_CTX_FMT_AARCH64)
+                               cb->tcr[1] |= TTBCR2_AS;
+               }
+       } else {
+               cb->tcr[0] = pgtbl_cfg->arm_lpae_s2_cfg.vtcr;
+       }
+
+       /* TTBRs */
+       if (stage1) {
+               if (cfg->fmt == ARM_SMMU_CTX_FMT_AARCH32_S) {
+                       cb->ttbr[0] = pgtbl_cfg->arm_v7s_cfg.ttbr[0];
+                       cb->ttbr[1] = pgtbl_cfg->arm_v7s_cfg.ttbr[1];
+               } else {
+                       cb->ttbr[0] = pgtbl_cfg->arm_lpae_s1_cfg.ttbr[0];
+                       cb->ttbr[0] |= (u64)cfg->asid << TTBRn_ASID_SHIFT;
+                       cb->ttbr[1] = pgtbl_cfg->arm_lpae_s1_cfg.ttbr[1];
+                       cb->ttbr[1] |= (u64)cfg->asid << TTBRn_ASID_SHIFT;
+               }
+       } else {
+               cb->ttbr[0] = pgtbl_cfg->arm_lpae_s2_cfg.vttbr;
+       }
+
+       /* MAIRs (stage-1 only) */
+       if (stage1) {
+               if (cfg->fmt == ARM_SMMU_CTX_FMT_AARCH32_S) {
+                       cb->mair[0] = pgtbl_cfg->arm_v7s_cfg.prrr;
+                       cb->mair[1] = pgtbl_cfg->arm_v7s_cfg.nmrr;
+               } else {
+                       cb->mair[0] = pgtbl_cfg->arm_lpae_s1_cfg.mair[0];
+                       cb->mair[1] = pgtbl_cfg->arm_lpae_s1_cfg.mair[1];
+               }
+       }
+}
+
+static void arm_smmu_write_context_bank(struct arm_smmu_device *smmu, int idx)
+{
+       u32 reg;
+       bool stage1;
+       struct arm_smmu_cb *cb = &smmu->cbs[idx];
+       struct arm_smmu_cfg *cfg = cb->cfg;
        void __iomem *cb_base, *gr1_base;
 
+       cb_base = ARM_SMMU_CB(smmu, idx);
+
+       /* Unassigned context banks only need disabling */
+       if (!cfg) {
+               writel_relaxed(0, cb_base + ARM_SMMU_CB_SCTLR);
+               return;
+       }
+
        gr1_base = ARM_SMMU_GR1(smmu);
        stage1 = cfg->cbar != CBAR_TYPE_S2_TRANS;
-       cb_base = ARM_SMMU_CB(smmu, cfg->cbndx);
 
+       /* CBA2R */
        if (smmu->version > ARM_SMMU_V1) {
                if (cfg->fmt == ARM_SMMU_CTX_FMT_AARCH64)
                        reg = CBA2R_RW64_64BIT;
@@ -796,7 +861,7 @@ static void arm_smmu_init_context_bank(struct arm_smmu_domain *smmu_domain,
                if (smmu->features & ARM_SMMU_FEAT_VMID16)
                        reg |= cfg->vmid << CBA2R_VMID_SHIFT;
 
-               writel_relaxed(reg, gr1_base + ARM_SMMU_GR1_CBA2R(cfg->cbndx));
+               writel_relaxed(reg, gr1_base + ARM_SMMU_GR1_CBA2R(idx));
        }
 
        /* CBAR */
@@ -815,72 +880,41 @@ static void arm_smmu_init_context_bank(struct arm_smmu_domain *smmu_domain,
                /* 8-bit VMIDs live in CBAR */
                reg |= cfg->vmid << CBAR_VMID_SHIFT;
        }
-       writel_relaxed(reg, gr1_base + ARM_SMMU_GR1_CBAR(cfg->cbndx));
+       writel_relaxed(reg, gr1_base + ARM_SMMU_GR1_CBAR(idx));
 
        /*
         * TTBCR
         * We must write this before the TTBRs, since it determines the
         * access behaviour of some fields (in particular, ASID[15:8]).
         */
-       if (stage1) {
-               if (cfg->fmt == ARM_SMMU_CTX_FMT_AARCH32_S) {
-                       reg = pgtbl_cfg->arm_v7s_cfg.tcr;
-                       reg2 = 0;
-               } else {
-                       reg = pgtbl_cfg->arm_lpae_s1_cfg.tcr;
-                       reg2 = pgtbl_cfg->arm_lpae_s1_cfg.tcr >> 32;
-                       reg2 |= TTBCR2_SEP_UPSTREAM;
-                       if (cfg->fmt == ARM_SMMU_CTX_FMT_AARCH64)
-                               reg2 |= TTBCR2_AS;
-               }
-               if (smmu->version > ARM_SMMU_V1)
-                       writel_relaxed(reg2, cb_base + ARM_SMMU_CB_TTBCR2);
-       } else {
-               reg = pgtbl_cfg->arm_lpae_s2_cfg.vtcr;
-       }
-       writel_relaxed(reg, cb_base + ARM_SMMU_CB_TTBCR);
+       if (stage1 && smmu->version > ARM_SMMU_V1)
+               writel_relaxed(cb->tcr[1], cb_base + ARM_SMMU_CB_TTBCR2);
+       writel_relaxed(cb->tcr[0], cb_base + ARM_SMMU_CB_TTBCR);
 
        /* TTBRs */
-       if (stage1) {
-               if (cfg->fmt == ARM_SMMU_CTX_FMT_AARCH32_S) {
-                       reg = pgtbl_cfg->arm_v7s_cfg.ttbr[0];
-                       writel_relaxed(reg, cb_base + ARM_SMMU_CB_TTBR0);
-                       reg = pgtbl_cfg->arm_v7s_cfg.ttbr[1];
-                       writel_relaxed(reg, cb_base + ARM_SMMU_CB_TTBR1);
-                       writel_relaxed(cfg->asid, cb_base + ARM_SMMU_CB_CONTEXTIDR);
-               } else {
-                       reg64 = pgtbl_cfg->arm_lpae_s1_cfg.ttbr[0];
-                       reg64 |= (u64)cfg->asid << TTBRn_ASID_SHIFT;
-                       writeq_relaxed(reg64, cb_base + ARM_SMMU_CB_TTBR0);
-                       reg64 = pgtbl_cfg->arm_lpae_s1_cfg.ttbr[1];
-                       reg64 |= (u64)cfg->asid << TTBRn_ASID_SHIFT;
-                       writeq_relaxed(reg64, cb_base + ARM_SMMU_CB_TTBR1);
-               }
+       if (cfg->fmt == ARM_SMMU_CTX_FMT_AARCH32_S) {
+               writel_relaxed(cfg->asid, cb_base + ARM_SMMU_CB_CONTEXTIDR);
+               writel_relaxed(cb->ttbr[0], cb_base + ARM_SMMU_CB_TTBR0);
+               writel_relaxed(cb->ttbr[1], cb_base + ARM_SMMU_CB_TTBR1);
        } else {
-               reg64 = pgtbl_cfg->arm_lpae_s2_cfg.vttbr;
-               writeq_relaxed(reg64, cb_base + ARM_SMMU_CB_TTBR0);
+               writeq_relaxed(cb->ttbr[0], cb_base + ARM_SMMU_CB_TTBR0);
+               if (stage1)
+                       writeq_relaxed(cb->ttbr[1], cb_base + ARM_SMMU_CB_TTBR1);
        }
 
        /* MAIRs (stage-1 only) */
        if (stage1) {
-               if (cfg->fmt == ARM_SMMU_CTX_FMT_AARCH32_S) {
-                       reg = pgtbl_cfg->arm_v7s_cfg.prrr;
-                       reg2 = pgtbl_cfg->arm_v7s_cfg.nmrr;
-               } else {
-                       reg = pgtbl_cfg->arm_lpae_s1_cfg.mair[0];
-                       reg2 = pgtbl_cfg->arm_lpae_s1_cfg.mair[1];
-               }
-               writel_relaxed(reg, cb_base + ARM_SMMU_CB_S1_MAIR0);
-               writel_relaxed(reg2, cb_base + ARM_SMMU_CB_S1_MAIR1);
+               writel_relaxed(cb->mair[0], cb_base + ARM_SMMU_CB_S1_MAIR0);
+               writel_relaxed(cb->mair[1], cb_base + ARM_SMMU_CB_S1_MAIR1);
        }
 
        /* SCTLR */
        reg = SCTLR_CFIE | SCTLR_CFRE | SCTLR_AFE | SCTLR_TRE | SCTLR_M;
        if (stage1)
                reg |= SCTLR_S1_ASIDPNE;
-#ifdef __BIG_ENDIAN
-       reg |= SCTLR_E;
-#endif
+       if (IS_ENABLED(CONFIG_CPU_BIG_ENDIAN))
+               reg |= SCTLR_E;
+
        writel_relaxed(reg, cb_base + ARM_SMMU_CB_SCTLR);
 }
 
@@ -1043,6 +1077,7 @@ static int arm_smmu_init_domain_context(struct iommu_domain *domain,
 
        /* Initialise the context bank with our page table cfg */
        arm_smmu_init_context_bank(smmu_domain, &pgtbl_cfg);
+       arm_smmu_write_context_bank(smmu, cfg->cbndx);
 
        /*
         * Request context fault interrupt. Do this last to avoid the
@@ -1075,7 +1110,6 @@ static void arm_smmu_destroy_domain_context(struct iommu_domain *domain)
        struct arm_smmu_domain *smmu_domain = to_smmu_domain(domain);
        struct arm_smmu_device *smmu = smmu_domain->smmu;
        struct arm_smmu_cfg *cfg = &smmu_domain->cfg;
-       void __iomem *cb_base;
        int irq;
 
        if (!smmu || domain->type == IOMMU_DOMAIN_IDENTITY)
@@ -1085,8 +1119,8 @@ static void arm_smmu_destroy_domain_context(struct iommu_domain *domain)
         * Disable the context bank and free the page tables before freeing
         * it.
         */
-       cb_base = ARM_SMMU_CB(smmu, cfg->cbndx);
-       writel_relaxed(0, cb_base + ARM_SMMU_CB_SCTLR);
+       smmu->cbs[cfg->cbndx].cfg = NULL;
+       arm_smmu_write_context_bank(smmu, cfg->cbndx);
 
        if (cfg->irptndx != INVALID_IRPTNDX) {
                irq = smmu->irqs[smmu->num_global_irqs + cfg->irptndx];
@@ -1729,7 +1763,6 @@ static struct iommu_ops arm_smmu_ops = {
 static void arm_smmu_device_reset(struct arm_smmu_device *smmu)
 {
        void __iomem *gr0_base = ARM_SMMU_GR0(smmu);
-       void __iomem *cb_base;
        int i;
        u32 reg, major;
 
@@ -1765,8 +1798,9 @@ static void arm_smmu_device_reset(struct arm_smmu_device *smmu)
 
        /* Make sure all context banks are disabled and clear CB_FSR  */
        for (i = 0; i < smmu->num_context_banks; ++i) {
-               cb_base = ARM_SMMU_CB(smmu, i);
-               writel_relaxed(0, cb_base + ARM_SMMU_CB_SCTLR);
+               void __iomem *cb_base = ARM_SMMU_CB(smmu, i);
+
+               arm_smmu_write_context_bank(smmu, i);
                writel_relaxed(FSR_FAULT, cb_base + ARM_SMMU_CB_FSR);
                /*
                 * Disable MMU-500's not-particularly-beneficial next-page
@@ -1972,6 +2006,10 @@ static int arm_smmu_device_cfg_probe(struct arm_smmu_device *smmu)
                smmu->cavium_id_base -= smmu->num_context_banks;
                dev_notice(smmu->dev, "\tenabling workaround for Cavium erratum 27704\n");
        }
+       smmu->cbs = devm_kcalloc(smmu->dev, smmu->num_context_banks,
+                                sizeof(*smmu->cbs), GFP_KERNEL);
+       if (!smmu->cbs)
+               return -ENOMEM;
 
        /* ID2 */
        id = readl_relaxed(gr0_base + ARM_SMMU_GR0_ID2);