From 748adbac7ca3452d0a1250c37ff4168d6ccf5ae0 Mon Sep 17 00:00:00 2001 From: sefidel Date: Tue, 14 Feb 2023 23:36:38 +0900 Subject: feat(nixos): init kompakt --- .../packages/linux-asahi/sven-iommu-4k.patch | 449 +++++++++++++++++++++ 1 file changed, 449 insertions(+) create mode 100644 nixos/modules/apple-silicon-support/packages/linux-asahi/sven-iommu-4k.patch (limited to 'nixos/modules/apple-silicon-support/packages/linux-asahi/sven-iommu-4k.patch') diff --git a/nixos/modules/apple-silicon-support/packages/linux-asahi/sven-iommu-4k.patch b/nixos/modules/apple-silicon-support/packages/linux-asahi/sven-iommu-4k.patch new file mode 100644 index 0000000..25eb541 --- /dev/null +++ b/nixos/modules/apple-silicon-support/packages/linux-asahi/sven-iommu-4k.patch @@ -0,0 +1,449 @@ +diff --git a/drivers/iommu/apple-dart.c b/drivers/iommu/apple-dart.c +index 4f1a37bdd42d..c8c3ea81d818 100644 +--- a/drivers/iommu/apple-dart.c ++++ b/drivers/iommu/apple-dart.c +@@ -97,7 +97,6 @@ struct apple_dart_hw { + * @lock: lock for hardware operations involving this dart + * @pgsize: pagesize supported by this DART + * @supports_bypass: indicates if this DART supports bypass mode +- * @force_bypass: force bypass mode due to pagesize mismatch? + * @sid2group: maps stream ids to iommu_groups + * @iommu: iommu core device + */ +@@ -115,7 +114,6 @@ struct apple_dart { + + u32 pgsize; + u32 supports_bypass : 1; +- u32 force_bypass : 1; + + struct iommu_group *sid2group[DART_MAX_STREAMS]; + struct iommu_device iommu; +@@ -499,9 +497,6 @@ static int apple_dart_attach_dev(struct iommu_domain *domain, + struct apple_dart_master_cfg *cfg = dev_iommu_priv_get(dev); + struct apple_dart_domain *dart_domain = to_dart_domain(domain); + +- if (cfg->stream_maps[0].dart->force_bypass && +- domain->type != IOMMU_DOMAIN_IDENTITY) +- return -EINVAL; + if (!cfg->stream_maps[0].dart->supports_bypass && + domain->type == IOMMU_DOMAIN_IDENTITY) + return -EINVAL; +@@ -630,8 +625,6 @@ static int apple_dart_of_xlate(struct device *dev, struct of_phandle_args *args) + if (cfg_dart) { + if (cfg_dart->supports_bypass != dart->supports_bypass) + return -EINVAL; +- if (cfg_dart->force_bypass != dart->force_bypass) +- return -EINVAL; + if (cfg_dart->pgsize != dart->pgsize) + return -EINVAL; + } +@@ -736,8 +729,6 @@ static int apple_dart_def_domain_type(struct device *dev) + { + struct apple_dart_master_cfg *cfg = dev_iommu_priv_get(dev); + +- if (cfg->stream_maps[0].dart->force_bypass) +- return IOMMU_DOMAIN_IDENTITY; + if (!cfg->stream_maps[0].dart->supports_bypass) + return IOMMU_DOMAIN_DMA; + +@@ -1121,8 +1121,6 @@ static int apple_dart_probe(struct platform_device *pdev) + goto err_clk_disable; + } + +- dart->force_bypass = dart->pgsize > PAGE_SIZE; +- + ret = apple_dart_hw_reset(dart); + if (ret) + goto err_clk_disable; +@@ -1149,8 +1147,8 @@ static int apple_dart_probe(struct platform_device *pdev) + + dev_info( + &pdev->dev, +- "DART [pagesize %x, %d streams, bypass support: %d, bypass forced: %d] initialized\n", +- dart->pgsize, dart->num_streams, dart->supports_bypass, dart->force_bypass); ++ "DART [pagesize %x, %d streams, bypass support: %d] initialized\n", ++ dart->pgsize, dart->num_streams, dart->supports_bypass); + return 0; + + err_sysfs_remove: + +diff --git a/drivers/iommu/dma-iommu.c b/drivers/iommu/dma-iommu.c +index 09f6e1c0f9c0..094592751cfa 100644 +--- a/drivers/iommu/dma-iommu.c ++++ b/drivers/iommu/dma-iommu.c +@@ -20,9 +20,11 @@ + #include + #include + #include ++#include + #include + #include + #include ++#include + #include + #include + #include +@@ -710,6 +712,9 @@ static struct page **__iommu_dma_alloc_pages(struct device *dev, + { + struct page **pages; + unsigned int i = 0, nid = dev_to_node(dev); ++ unsigned int j; ++ unsigned long min_order = __fls(order_mask); ++ unsigned int min_order_size = 1U << min_order; + + order_mask &= (2U << MAX_ORDER) - 1; + if (!order_mask) +@@ -749,15 +754,37 @@ static struct page **__iommu_dma_alloc_pages(struct device *dev, + split_page(page, order); + break; + } +- if (!page) { +- __iommu_dma_free_pages(pages, i); +- return NULL; ++ ++ /* ++ * If we have no valid page here we might be trying to allocate ++ * the last block consisting of 1<pgsize_bitmap; ++ struct sg_append_table sgt_append = {}; ++ struct scatterlist *last_sg; + struct page **pages; + dma_addr_t iova; + ssize_t ret; ++ phys_addr_t orig_s_phys; ++ size_t orig_s_len, orig_s_off, s_iova_off, iova_size; + + if (static_branch_unlikely(&iommu_deferred_attach_enabled) && + iommu_deferred_attach(dev, domain)) + return NULL; + + min_size = alloc_sizes & -alloc_sizes; +- if (min_size < PAGE_SIZE) { ++ if (iovad->granule > PAGE_SIZE) { ++ if (size < iovad->granule) { ++ /* ensure a single contiguous allocation */ ++ min_size = ALIGN(size, PAGE_SIZE*(1U<coherent_dma_mask, dev); ++ iova_size = iova_align(iovad, size); ++ iova = iommu_dma_alloc_iova(domain, iova_size, dev->coherent_dma_mask, dev); + if (!iova) + goto out_free_pages; + +- if (sg_alloc_table_from_pages(sgt, pages, count, 0, size, GFP_KERNEL)) ++ /* append_table is only used to get a pointer to the last entry */ ++ if (sg_alloc_append_table_from_pages(&sgt_append, pages, count, 0, ++ iova_size, UINT_MAX, 0, GFP_KERNEL)) + goto out_free_iova; ++ memcpy(sgt, &sgt_append.sgt, sizeof(*sgt)); ++ last_sg = sgt_append.prv; + + if (!(ioprot & IOMMU_CACHE)) { + struct scatterlist *sg; +@@ -825,18 +839,59 @@ static struct page **__iommu_dma_alloc_noncontiguous(struct device *dev, + arch_dma_prep_coherent(sg_page(sg), sg->length); + } + ++ if (iovad->granule > PAGE_SIZE) { ++ if (size < iovad->granule) { ++ /* ++ * we only have a single sg list entry here that is ++ * likely not aligned to iovad->granule. adjust the ++ * entry to represent the encapsulating IOMMU page ++ * and then later restore everything to its original ++ * values, similar to the impedance matching done in ++ * iommu_dma_map_sg. ++ */ ++ orig_s_phys = sg_phys(sgt->sgl); ++ orig_s_len = sgt->sgl->length; ++ orig_s_off = sgt->sgl->offset; ++ s_iova_off = iova_offset(iovad, orig_s_phys); ++ ++ sg_set_page(sgt->sgl, ++ pfn_to_page(PHYS_PFN(orig_s_phys - s_iova_off)), ++ iova_align(iovad, orig_s_len + s_iova_off), ++ sgt->sgl->offset & ~s_iova_off); ++ } else { ++ /* ++ * convince iommu_map_sg_atomic to map the last block ++ * even though it may be too small. ++ */ ++ orig_s_len = last_sg->length; ++ last_sg->length = iova_align(iovad, last_sg->length); ++ } ++ } ++ + ret = iommu_map_sg_atomic(domain, iova, sgt->sgl, sgt->orig_nents, ioprot); +- if (ret < 0 || ret < size) ++ if (ret < 0 || ret < iova_size) + goto out_free_sg; + ++ if (iovad->granule > PAGE_SIZE) { ++ if (size < iovad->granule) { ++ sg_set_page(sgt->sgl, ++ pfn_to_page(PHYS_PFN(orig_s_phys)), ++ orig_s_len, orig_s_off); ++ ++ iova += s_iova_off; ++ } else { ++ last_sg->length = orig_s_len; ++ } ++ } ++ + sgt->sgl->dma_address = iova; +- sgt->sgl->dma_length = size; ++ sgt->sgl->dma_length = iova_size; + return pages; + + out_free_sg: + sg_free_table(sgt); + out_free_iova: +- iommu_dma_free_iova(cookie, iova, size, NULL); ++ iommu_dma_free_iova(cookie, iova, iova_size, NULL); + out_free_pages: + __iommu_dma_free_pages(pages, count); + return NULL; +@@ -1040,8 +1124,9 @@ static int __finalise_sg(struct device *dev, struct scatterlist *sg, int nents, + unsigned int s_length = sg_dma_len(s); + unsigned int s_iova_len = s->length; + +- s->offset += s_iova_off; +- s->length = s_length; ++ sg_set_page(s, ++ pfn_to_page(PHYS_PFN(sg_phys(s) + s_iova_off)), ++ s_length, s_iova_off & ~PAGE_MASK); + sg_dma_address(s) = DMA_MAPPING_ERROR; + sg_dma_len(s) = 0; + +@@ -1082,13 +1167,17 @@ static int __finalise_sg(struct device *dev, struct scatterlist *sg, int nents, + static void __invalidate_sg(struct scatterlist *sg, int nents) + { + struct scatterlist *s; ++ phys_addr_t orig_paddr; + int i; + + for_each_sg(sg, s, nents, i) { +- if (sg_dma_address(s) != DMA_MAPPING_ERROR) +- s->offset += sg_dma_address(s); +- if (sg_dma_len(s)) +- s->length = sg_dma_len(s); ++ if (sg_dma_len(s)) { ++ orig_paddr = sg_phys(s) + sg_dma_address(s); ++ sg_set_page(s, ++ pfn_to_page(PHYS_PFN(orig_paddr)), ++ sg_dma_len(s), ++ sg_dma_address(s) & ~PAGE_MASK); ++ } + sg_dma_address(s) = DMA_MAPPING_ERROR; + sg_dma_len(s) = 0; + } +@@ -1166,15 +1255,16 @@ static int iommu_dma_map_sg(struct device *dev, struct scatterlist *sg, + * stashing the unaligned parts in the as-yet-unused DMA fields. + */ + for_each_sg(sg, s, nents, i) { +- size_t s_iova_off = iova_offset(iovad, s->offset); ++ phys_addr_t s_phys = sg_phys(s); ++ size_t s_iova_off = iova_offset(iovad, s_phys); + size_t s_length = s->length; + size_t pad_len = (mask - iova_len + 1) & mask; + + sg_dma_address(s) = s_iova_off; + sg_dma_len(s) = s_length; +- s->offset -= s_iova_off; + s_length = iova_align(iovad, s_length + s_iova_off); +- s->length = s_length; ++ sg_set_page(s, pfn_to_page(PHYS_PFN(s_phys - s_iova_off)), ++ s_length, s->offset & ~s_iova_off); + + /* + * Due to the alignment of our single IOVA allocation, we can +@@ -1412,9 +1502,15 @@ static int iommu_dma_get_sgtable(struct device *dev, struct sg_table *sgt, + void *cpu_addr, dma_addr_t dma_addr, size_t size, + unsigned long attrs) + { ++ struct iommu_domain *domain = iommu_get_dma_domain(dev); ++ struct iommu_dma_cookie *cookie = domain->iova_cookie; ++ struct iova_domain *iovad = &cookie->iovad; + struct page *page; + int ret; + ++ if (iovad->granule > PAGE_SIZE) ++ return -ENXIO; ++ + if (is_vmalloc_addr(cpu_addr)) { + struct page **pages = dma_common_find_pages(cpu_addr); + +diff --git a/drivers/iommu/iommu.c b/drivers/iommu/iommu.c +index f2c45b85b9fc..0c370e486d6e 100644 +--- a/drivers/iommu/iommu.c ++++ b/drivers/iommu/iommu.c +@@ -80,6 +80,8 @@ static struct iommu_domain *__iommu_domain_alloc(struct bus_type *bus, + unsigned type); + static int __iommu_attach_device(struct iommu_domain *domain, + struct device *dev); ++static void __iommu_detach_device(struct iommu_domain *domain, ++ struct device *dev); + static int __iommu_attach_group(struct iommu_domain *domain, + struct iommu_group *group); + static void __iommu_detach_group(struct iommu_domain *domain, +@@ -1976,6 +1978,24 @@ void iommu_domain_free(struct iommu_domain *domain) + } + EXPORT_SYMBOL_GPL(iommu_domain_free); + ++static int iommu_check_page_size(struct iommu_domain *domain, ++ struct device *dev) ++{ ++ bool trusted = !(dev_is_pci(dev) && to_pci_dev(dev)->untrusted); ++ ++ if (!iommu_is_paging_domain(domain)) ++ return 0; ++ if (iommu_is_large_pages_domain(domain) && trusted) ++ return 0; ++ ++ if (!(domain->pgsize_bitmap & (PAGE_SIZE | (PAGE_SIZE - 1)))) { ++ pr_warn("IOMMU pages cannot exactly represent CPU pages.\n"); ++ return -EFAULT; ++ } ++ ++ return 0; ++} ++ + static int __iommu_attach_device(struct iommu_domain *domain, + struct device *dev) + { +@@ -1985,9 +2005,23 @@ static int __iommu_attach_device(struct iommu_domain *domain, + return -ENODEV; + + ret = domain->ops->attach_dev(domain, dev); +- if (!ret) +- trace_attach_device_to_domain(dev); +- return ret; ++ if (ret) ++ return ret; ++ ++ /* ++ * Check that CPU pages can be represented by the IOVA granularity. ++ * This has to be done after ops->attach_dev since many IOMMU drivers ++ * only limit domain->pgsize_bitmap after having attached the first ++ * device. ++ */ ++ ret = iommu_check_page_size(domain, dev); ++ if (ret) { ++ __iommu_detach_device(domain, dev); ++ return ret; ++ } ++ ++ trace_attach_device_to_domain(dev); ++ return 0; + } + + int iommu_attach_device(struct iommu_domain *domain, struct device *dev) +diff --git a/drivers/iommu/iova.c b/drivers/iommu/iova.c +index db77aa675145..180ce65a6789 100644 +--- a/drivers/iommu/iova.c ++++ b/drivers/iommu/iova.c +@@ -49,10 +49,11 @@ init_iova_domain(struct iova_domain *iovad, unsigned long granule, + { + /* + * IOVA granularity will normally be equal to the smallest +- * supported IOMMU page size; both *must* be capable of +- * representing individual CPU pages exactly. ++ * supported IOMMU page size; while both usually are capable of ++ * representing individual CPU pages exactly the IOVA allocator ++ * supports any granularities that are an exact power of two. + */ +- BUG_ON((granule > PAGE_SIZE) || !is_power_of_2(granule)); ++ BUG_ON(!is_power_of_2(granule)); + + spin_lock_init(&iovad->iova_rbtree_lock); + iovad->rbroot = RB_ROOT; +diff --git a/include/linux/iommu.h b/include/linux/iommu.h +index 9208eca4b0d1..dec2dd70a876 100644 +--- a/include/linux/iommu.h ++++ b/include/linux/iommu.h +@@ -63,6 +63,8 @@ struct iommu_domain_geometry { + implementation */ + #define __IOMMU_DOMAIN_PT (1U << 2) /* Domain is identity mapped */ + #define __IOMMU_DOMAIN_DMA_FQ (1U << 3) /* DMA-API uses flush queue */ ++#define __IOMMU_DOMAIN_LP (1U << 4) /* Support for PAGE_SIZE smaller ++ than IOMMU page size */ + + /* + * This are the possible domain-types +@@ -82,10 +84,12 @@ struct iommu_domain_geometry { + #define IOMMU_DOMAIN_IDENTITY (__IOMMU_DOMAIN_PT) + #define IOMMU_DOMAIN_UNMANAGED (__IOMMU_DOMAIN_PAGING) + #define IOMMU_DOMAIN_DMA (__IOMMU_DOMAIN_PAGING | \ +- __IOMMU_DOMAIN_DMA_API) ++ __IOMMU_DOMAIN_DMA_API | \ ++ __IOMMU_DOMAIN_LP) + #define IOMMU_DOMAIN_DMA_FQ (__IOMMU_DOMAIN_PAGING | \ + __IOMMU_DOMAIN_DMA_API | \ +- __IOMMU_DOMAIN_DMA_FQ) ++ __IOMMU_DOMAIN_DMA_FQ | \ ++ __IOMMU_DOMAIN_LP) + + struct iommu_domain { + unsigned type; +@@ -102,6 +106,16 @@ static inline bool iommu_is_dma_domain(struct iommu_domain *domain) + return domain->type & __IOMMU_DOMAIN_DMA_API; + } + ++static inline bool iommu_is_paging_domain(struct iommu_domain *domain) ++{ ++ return domain->type & __IOMMU_DOMAIN_PAGING; ++} ++ ++static inline bool iommu_is_large_pages_domain(struct iommu_domain *domain) ++{ ++ return domain->type & __IOMMU_DOMAIN_LP; ++} ++ + enum iommu_cap { + IOMMU_CAP_CACHE_COHERENCY, /* IOMMU can enforce cache coherent DMA + transactions */ -- cgit 1.4.1