From: Fangyu Yu Replace .domain_alloc_paging with .domain_alloc_paging_flags so callers can pass allocation flags to select the appropriate page-table type. When IOMMU_HWPT_ALLOC_NEST_PARENT or IOMMU_HWPT_ALLOC_DIRTY_TRACKING is set in @flags, allocate a second-stage (iohgatp) domain. When @flags is 0 the behaviour is identical to the previous domain_alloc_paging: first-stage (iosatp) domain. Signed-off-by: Fangyu Yu --- drivers/iommu/riscv/iommu.c | 90 +++++++++++++++++++++++++++---------- 1 file changed, 67 insertions(+), 23 deletions(-) diff --git a/drivers/iommu/riscv/iommu.c b/drivers/iommu/riscv/iommu.c index e883ace2f4f1..ebf42f74e194 100644 --- a/drivers/iommu/riscv/iommu.c +++ b/drivers/iommu/riscv/iommu.c @@ -1255,25 +1255,21 @@ static const struct iommu_domain_ops riscv_iommu_paging_domain_ops = { .flush_iotlb_all = riscv_iommu_iotlb_flush_all, }; -static struct iommu_domain *riscv_iommu_alloc_paging_domain(struct device *dev) +static struct iommu_domain *riscv_iommu_domain_alloc_paging_flags( + struct device *dev, u32 flags, + const struct iommu_user_data *user_data) { struct pt_iommu_riscv_64_cfg cfg = {}; struct riscv_iommu_domain *domain; struct riscv_iommu_device *iommu; int ret; + const u32 supported_flags = IOMMU_HWPT_ALLOC_DIRTY_TRACKING | + IOMMU_HWPT_ALLOC_NEST_PARENT; - iommu = dev_to_iommu(dev); - if (iommu->caps & RISCV_IOMMU_CAPABILITIES_SV57) { - cfg.common.hw_max_vasz_lg2 = 57; - } else if (iommu->caps & RISCV_IOMMU_CAPABILITIES_SV48) { - cfg.common.hw_max_vasz_lg2 = 48; - } else if (iommu->caps & RISCV_IOMMU_CAPABILITIES_SV39) { - cfg.common.hw_max_vasz_lg2 = 39; - } else { - dev_err(dev, "cannot find supported page table mode\n"); - return ERR_PTR(-ENODEV); - } - cfg.common.hw_max_oasz_lg2 = 56; + if (flags & ~supported_flags) + return ERR_PTR(-EOPNOTSUPP); + if (user_data) + return ERR_PTR(-EOPNOTSUPP); domain = kzalloc_obj(*domain); if (!domain) @@ -1281,6 +1277,8 @@ static struct iommu_domain *riscv_iommu_alloc_paging_domain(struct device *dev) INIT_LIST_HEAD_RCU(&domain->bonds); spin_lock_init(&domain->lock); + iommu = dev_to_iommu(dev); + cfg.common.hw_max_oasz_lg2 = 56; /* * 6.4 IOMMU capabilities [..] IOMMU implementations must support the * Svnapot standard extension for NAPOT Translation Contiguity. @@ -1291,19 +1289,65 @@ static struct iommu_domain *riscv_iommu_alloc_paging_domain(struct device *dev) domain->riscvpt.iommu.nid = dev_to_node(iommu->dev); domain->domain.ops = &riscv_iommu_paging_domain_ops; - domain->pscid = ida_alloc_range(&riscv_iommu_pscids, 1, - RISCV_IOMMU_MAX_PSCID, GFP_KERNEL); - if (domain->pscid < 0) { - riscv_iommu_free_paging_domain(&domain->domain); - return ERR_PTR(-ENOMEM); + switch (flags) { + case 0: + if (iommu->caps & RISCV_IOMMU_CAPABILITIES_SV57) { + cfg.common.hw_max_vasz_lg2 = 57; + } else if (iommu->caps & RISCV_IOMMU_CAPABILITIES_SV48) { + cfg.common.hw_max_vasz_lg2 = 48; + } else if (iommu->caps & RISCV_IOMMU_CAPABILITIES_SV39) { + cfg.common.hw_max_vasz_lg2 = 39; + } else { + ret = -ENODEV; + goto err_free; + } + domain->pscid = ida_alloc_range(&riscv_iommu_pscids, 1, + RISCV_IOMMU_MAX_PSCID, GFP_KERNEL); + if (domain->pscid < 0) { + ret = -ENOMEM; + goto err_free; + } + break; + case IOMMU_HWPT_ALLOC_NEST_PARENT: + case IOMMU_HWPT_ALLOC_DIRTY_TRACKING: + case IOMMU_HWPT_ALLOC_DIRTY_TRACKING | IOMMU_HWPT_ALLOC_NEST_PARENT: + /* + * Second-stage (iohgatp) page table for KVM VFIO device + * pass-through and dirty tracking. The GPA space is 2 bits + * wider than the corresponding first-stage VA space (x4 root + * page table), so hw_max_vasz_lg2 values are 41/50/59. + */ + if (iommu->caps & RISCV_IOMMU_CAPABILITIES_SV57X4) { + cfg.common.hw_max_vasz_lg2 = 59; + } else if (iommu->caps & RISCV_IOMMU_CAPABILITIES_SV48X4) { + cfg.common.hw_max_vasz_lg2 = 50; + } else if (iommu->caps & RISCV_IOMMU_CAPABILITIES_SV39X4) { + cfg.common.hw_max_vasz_lg2 = 41; + } else { + ret = -ENODEV; + goto err_free; + } + domain->gscid = ida_alloc_range(&riscv_iommu_gscids, 1, + RISCV_IOMMU_MAX_GSCID, GFP_KERNEL); + if (domain->gscid < 0) { + ret = -ENOMEM; + goto err_free; + } + cfg.common.features |= BIT(PT_FEAT_RISCV_S2); + break; + default: + ret = -EOPNOTSUPP; + goto err_free; } ret = pt_iommu_riscv_64_init(&domain->riscvpt, &cfg, GFP_KERNEL); - if (ret) { - riscv_iommu_free_paging_domain(&domain->domain); - return ERR_PTR(ret); - } + if (ret) + goto err_free; return &domain->domain; + +err_free: + riscv_iommu_free_paging_domain(&domain->domain); + return ERR_PTR(ret); } static int riscv_iommu_attach_blocking_domain(struct iommu_domain *iommu_domain, @@ -1438,7 +1482,7 @@ static const struct iommu_ops riscv_iommu_ops = { .identity_domain = &riscv_iommu_identity_domain, .blocked_domain = &riscv_iommu_blocking_domain, .release_domain = &riscv_iommu_blocking_domain, - .domain_alloc_paging = riscv_iommu_alloc_paging_domain, + .domain_alloc_paging_flags = riscv_iommu_domain_alloc_paging_flags, .device_group = riscv_iommu_device_group, .probe_device = riscv_iommu_probe_device, .release_device = riscv_iommu_release_device, -- 2.50.1