提交 c55b7a34 编写于 作者: K Kunkun Jiang 提交者: Zheng Zengkai

Revert "iommu/smmuv3: Allow stage 1 invalidation with unmanaged ASIDs"

virt inclusion
category: bugfix
bugzilla: https://gitee.com/openeuler/kernel/issues/I61SPO
CVE: NA

--------------------------------

This reverts commit eeb79c56.
Signed-off-by: NKunkun Jiang <jiangkunkun@huawei.com>
Reviewed-by: NKeqian Zhu <zhukeqian1@huawei.com>
Signed-off-by: NZheng Zengkai <zhengzengkai@huawei.com>
上级 3a1cc4f0
...@@ -2191,9 +2191,9 @@ int arm_smmu_atc_inv_domain(struct arm_smmu_domain *smmu_domain, int ssid, ...@@ -2191,9 +2191,9 @@ int arm_smmu_atc_inv_domain(struct arm_smmu_domain *smmu_domain, int ssid,
} }
/* IO_PGTABLE API */ /* IO_PGTABLE API */
static void __arm_smmu_tlb_inv_context(struct arm_smmu_domain *smmu_domain, static void arm_smmu_tlb_inv_context(void *cookie)
int ext_asid)
{ {
struct arm_smmu_domain *smmu_domain = cookie;
struct arm_smmu_device *smmu = smmu_domain->smmu; struct arm_smmu_device *smmu = smmu_domain->smmu;
struct arm_smmu_cmdq_ent cmd; struct arm_smmu_cmdq_ent cmd;
...@@ -2204,12 +2204,7 @@ static void __arm_smmu_tlb_inv_context(struct arm_smmu_domain *smmu_domain, ...@@ -2204,12 +2204,7 @@ static void __arm_smmu_tlb_inv_context(struct arm_smmu_domain *smmu_domain,
* insertion to guarantee those are observed before the TLBI. Do be * insertion to guarantee those are observed before the TLBI. Do be
* careful, 007. * careful, 007.
*/ */
if (ext_asid >= 0) { /* guest stage 1 invalidation */ if (smmu_domain->stage == ARM_SMMU_DOMAIN_S1) {
cmd.opcode = CMDQ_OP_TLBI_NH_ASID;
cmd.tlbi.asid = ext_asid;
cmd.tlbi.vmid = smmu_domain->s2_cfg.vmid;
arm_smmu_cmdq_issue_cmd_with_sync(smmu, &cmd);
} else if (smmu_domain->stage == ARM_SMMU_DOMAIN_S1) {
arm_smmu_tlb_inv_asid(smmu, smmu_domain->s1_cfg.cd.asid); arm_smmu_tlb_inv_asid(smmu, smmu_domain->s1_cfg.cd.asid);
} else { } else {
cmd.opcode = CMDQ_OP_TLBI_S12_VMALL; cmd.opcode = CMDQ_OP_TLBI_S12_VMALL;
...@@ -2224,13 +2219,6 @@ static void __arm_smmu_tlb_inv_context(struct arm_smmu_domain *smmu_domain, ...@@ -2224,13 +2219,6 @@ static void __arm_smmu_tlb_inv_context(struct arm_smmu_domain *smmu_domain,
} }
static void arm_smmu_tlb_inv_context(void *cookie)
{
struct arm_smmu_domain *smmu_domain = cookie;
__arm_smmu_tlb_inv_context(smmu_domain, -1);
}
static void __arm_smmu_tlb_inv_range(struct arm_smmu_cmdq_ent *cmd, static void __arm_smmu_tlb_inv_range(struct arm_smmu_cmdq_ent *cmd,
unsigned long iova, size_t size, unsigned long iova, size_t size,
size_t granule, size_t granule,
...@@ -2292,10 +2280,9 @@ static void __arm_smmu_tlb_inv_range(struct arm_smmu_cmdq_ent *cmd, ...@@ -2292,10 +2280,9 @@ static void __arm_smmu_tlb_inv_range(struct arm_smmu_cmdq_ent *cmd,
arm_smmu_preempt_enable(smmu); arm_smmu_preempt_enable(smmu);
} }
static void static void arm_smmu_tlb_inv_range_domain(unsigned long iova, size_t size,
arm_smmu_tlb_inv_range_domain(unsigned long iova, size_t size, size_t granule, bool leaf,
size_t granule, bool leaf, int ext_asid, struct arm_smmu_domain *smmu_domain)
struct arm_smmu_domain *smmu_domain)
{ {
struct arm_smmu_cmdq_ent cmd = { struct arm_smmu_cmdq_ent cmd = {
.tlbi = { .tlbi = {
...@@ -2303,16 +2290,7 @@ arm_smmu_tlb_inv_range_domain(unsigned long iova, size_t size, ...@@ -2303,16 +2290,7 @@ arm_smmu_tlb_inv_range_domain(unsigned long iova, size_t size,
}, },
}; };
if (ext_asid >= 0) { /* guest stage 1 invalidation */ if (smmu_domain->stage == ARM_SMMU_DOMAIN_S1) {
/*
* At the moment the guest only uses NS-EL1, to be
* revisited when nested virt gets supported with E2H
* exposed.
*/
cmd.opcode = CMDQ_OP_TLBI_NH_VA;
cmd.tlbi.asid = ext_asid;
cmd.tlbi.vmid = smmu_domain->s2_cfg.vmid;
} else if (smmu_domain->stage == ARM_SMMU_DOMAIN_S1) {
cmd.opcode = smmu_domain->smmu->features & ARM_SMMU_FEAT_E2H ? cmd.opcode = smmu_domain->smmu->features & ARM_SMMU_FEAT_E2H ?
CMDQ_OP_TLBI_EL2_VA : CMDQ_OP_TLBI_NH_VA; CMDQ_OP_TLBI_EL2_VA : CMDQ_OP_TLBI_NH_VA;
cmd.tlbi.asid = smmu_domain->s1_cfg.cd.asid; cmd.tlbi.asid = smmu_domain->s1_cfg.cd.asid;
...@@ -2320,7 +2298,6 @@ arm_smmu_tlb_inv_range_domain(unsigned long iova, size_t size, ...@@ -2320,7 +2298,6 @@ arm_smmu_tlb_inv_range_domain(unsigned long iova, size_t size,
cmd.opcode = CMDQ_OP_TLBI_S2_IPA; cmd.opcode = CMDQ_OP_TLBI_S2_IPA;
cmd.tlbi.vmid = smmu_domain->s2_cfg.vmid; cmd.tlbi.vmid = smmu_domain->s2_cfg.vmid;
} }
__arm_smmu_tlb_inv_range(&cmd, iova, size, granule, smmu_domain); __arm_smmu_tlb_inv_range(&cmd, iova, size, granule, smmu_domain);
/* /*
...@@ -2363,7 +2340,7 @@ static void arm_smmu_tlb_inv_page_nosync(struct iommu_iotlb_gather *gather, ...@@ -2363,7 +2340,7 @@ static void arm_smmu_tlb_inv_page_nosync(struct iommu_iotlb_gather *gather,
static void arm_smmu_tlb_inv_walk(unsigned long iova, size_t size, static void arm_smmu_tlb_inv_walk(unsigned long iova, size_t size,
size_t granule, void *cookie) size_t granule, void *cookie)
{ {
arm_smmu_tlb_inv_range_domain(iova, size, granule, false, -1, cookie); arm_smmu_tlb_inv_range_domain(iova, size, granule, false, cookie);
} }
static const struct iommu_flush_ops arm_smmu_flush_ops = { static const struct iommu_flush_ops arm_smmu_flush_ops = {
...@@ -2999,9 +2976,8 @@ static void arm_smmu_iotlb_sync(struct iommu_domain *domain, ...@@ -2999,9 +2976,8 @@ static void arm_smmu_iotlb_sync(struct iommu_domain *domain,
{ {
struct arm_smmu_domain *smmu_domain = to_smmu_domain(domain); struct arm_smmu_domain *smmu_domain = to_smmu_domain(domain);
arm_smmu_tlb_inv_range_domain(gather->start, arm_smmu_tlb_inv_range_domain(gather->start, gather->end - gather->start + 1,
gather->end - gather->start + 1, gather->pgsize, true, smmu_domain);
gather->pgsize, true, -1, smmu_domain);
} }
static phys_addr_t static phys_addr_t
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册