From 08d7136a2e3150dd898cf2d2d3756801dfb407bd Mon Sep 17 00:00:00 2001 From: sefidel Date: Wed, 15 Feb 2023 14:32:53 +0900 Subject: feat(nixos/kompakt): use flakes for asahi-related dependencies Imports are done directly inside the system configuration file to ensure one doesn't forget to import the required modules. --- .../packages/linux-asahi/sven-iommu-4k.patch | 449 --------------------- 1 file changed, 449 deletions(-) delete mode 100644 nixos/modules/apple-silicon-support/packages/linux-asahi/sven-iommu-4k.patch (limited to 'nixos/modules/apple-silicon-support/packages/linux-asahi/sven-iommu-4k.patch') diff --git a/nixos/modules/apple-silicon-support/packages/linux-asahi/sven-iommu-4k.patch b/nixos/modules/apple-silicon-support/packages/linux-asahi/sven-iommu-4k.patch deleted file mode 100644 index 25eb541..0000000 --- a/nixos/modules/apple-silicon-support/packages/linux-asahi/sven-iommu-4k.patch +++ /dev/null @@ -1,449 +0,0 @@ -diff --git a/drivers/iommu/apple-dart.c b/drivers/iommu/apple-dart.c -index 4f1a37bdd42d..c8c3ea81d818 100644 ---- a/drivers/iommu/apple-dart.c -+++ b/drivers/iommu/apple-dart.c -@@ -97,7 +97,6 @@ struct apple_dart_hw { - * @lock: lock for hardware operations involving this dart - * @pgsize: pagesize supported by this DART - * @supports_bypass: indicates if this DART supports bypass mode -- * @force_bypass: force bypass mode due to pagesize mismatch? - * @sid2group: maps stream ids to iommu_groups - * @iommu: iommu core device - */ -@@ -115,7 +114,6 @@ struct apple_dart { - - u32 pgsize; - u32 supports_bypass : 1; -- u32 force_bypass : 1; - - struct iommu_group *sid2group[DART_MAX_STREAMS]; - struct iommu_device iommu; -@@ -499,9 +497,6 @@ static int apple_dart_attach_dev(struct iommu_domain *domain, - struct apple_dart_master_cfg *cfg = dev_iommu_priv_get(dev); - struct apple_dart_domain *dart_domain = to_dart_domain(domain); - -- if (cfg->stream_maps[0].dart->force_bypass && -- domain->type != IOMMU_DOMAIN_IDENTITY) -- return -EINVAL; - if (!cfg->stream_maps[0].dart->supports_bypass && - domain->type == IOMMU_DOMAIN_IDENTITY) - return -EINVAL; -@@ -630,8 +625,6 @@ static int apple_dart_of_xlate(struct device *dev, struct of_phandle_args *args) - if (cfg_dart) { - if (cfg_dart->supports_bypass != dart->supports_bypass) - return -EINVAL; -- if (cfg_dart->force_bypass != dart->force_bypass) -- return -EINVAL; - if (cfg_dart->pgsize != dart->pgsize) - return -EINVAL; - } -@@ -736,8 +729,6 @@ static int apple_dart_def_domain_type(struct device *dev) - { - struct apple_dart_master_cfg *cfg = dev_iommu_priv_get(dev); - -- if (cfg->stream_maps[0].dart->force_bypass) -- return IOMMU_DOMAIN_IDENTITY; - if (!cfg->stream_maps[0].dart->supports_bypass) - return IOMMU_DOMAIN_DMA; - -@@ -1121,8 +1121,6 @@ static int apple_dart_probe(struct platform_device *pdev) - goto err_clk_disable; - } - -- dart->force_bypass = dart->pgsize > PAGE_SIZE; -- - ret = apple_dart_hw_reset(dart); - if (ret) - goto err_clk_disable; -@@ -1149,8 +1147,8 @@ static int apple_dart_probe(struct platform_device *pdev) - - dev_info( - &pdev->dev, -- "DART [pagesize %x, %d streams, bypass support: %d, bypass forced: %d] initialized\n", -- dart->pgsize, dart->num_streams, dart->supports_bypass, dart->force_bypass); -+ "DART [pagesize %x, %d streams, bypass support: %d] initialized\n", -+ dart->pgsize, dart->num_streams, dart->supports_bypass); - return 0; - - err_sysfs_remove: - -diff --git a/drivers/iommu/dma-iommu.c b/drivers/iommu/dma-iommu.c -index 09f6e1c0f9c0..094592751cfa 100644 ---- a/drivers/iommu/dma-iommu.c -+++ b/drivers/iommu/dma-iommu.c -@@ -20,9 +20,11 @@ - #include - #include - #include -+#include - #include - #include - #include -+#include - #include - #include - #include -@@ -710,6 +712,9 @@ static struct page **__iommu_dma_alloc_pages(struct device *dev, - { - struct page **pages; - unsigned int i = 0, nid = dev_to_node(dev); -+ unsigned int j; -+ unsigned long min_order = __fls(order_mask); -+ unsigned int min_order_size = 1U << min_order; - - order_mask &= (2U << MAX_ORDER) - 1; - if (!order_mask) -@@ -749,15 +754,37 @@ static struct page **__iommu_dma_alloc_pages(struct device *dev, - split_page(page, order); - break; - } -- if (!page) { -- __iommu_dma_free_pages(pages, i); -- return NULL; -+ -+ /* -+ * If we have no valid page here we might be trying to allocate -+ * the last block consisting of 1<pgsize_bitmap; -+ struct sg_append_table sgt_append = {}; -+ struct scatterlist *last_sg; - struct page **pages; - dma_addr_t iova; - ssize_t ret; -+ phys_addr_t orig_s_phys; -+ size_t orig_s_len, orig_s_off, s_iova_off, iova_size; - - if (static_branch_unlikely(&iommu_deferred_attach_enabled) && - iommu_deferred_attach(dev, domain)) - return NULL; - - min_size = alloc_sizes & -alloc_sizes; -- if (min_size < PAGE_SIZE) { -+ if (iovad->granule > PAGE_SIZE) { -+ if (size < iovad->granule) { -+ /* ensure a single contiguous allocation */ -+ min_size = ALIGN(size, PAGE_SIZE*(1U<coherent_dma_mask, dev); -+ iova_size = iova_align(iovad, size); -+ iova = iommu_dma_alloc_iova(domain, iova_size, dev->coherent_dma_mask, dev); - if (!iova) - goto out_free_pages; - -- if (sg_alloc_table_from_pages(sgt, pages, count, 0, size, GFP_KERNEL)) -+ /* append_table is only used to get a pointer to the last entry */ -+ if (sg_alloc_append_table_from_pages(&sgt_append, pages, count, 0, -+ iova_size, UINT_MAX, 0, GFP_KERNEL)) - goto out_free_iova; -+ memcpy(sgt, &sgt_append.sgt, sizeof(*sgt)); -+ last_sg = sgt_append.prv; - - if (!(ioprot & IOMMU_CACHE)) { - struct scatterlist *sg; -@@ -825,18 +839,59 @@ static struct page **__iommu_dma_alloc_noncontiguous(struct device *dev, - arch_dma_prep_coherent(sg_page(sg), sg->length); - } - -+ if (iovad->granule > PAGE_SIZE) { -+ if (size < iovad->granule) { -+ /* -+ * we only have a single sg list entry here that is -+ * likely not aligned to iovad->granule. adjust the -+ * entry to represent the encapsulating IOMMU page -+ * and then later restore everything to its original -+ * values, similar to the impedance matching done in -+ * iommu_dma_map_sg. -+ */ -+ orig_s_phys = sg_phys(sgt->sgl); -+ orig_s_len = sgt->sgl->length; -+ orig_s_off = sgt->sgl->offset; -+ s_iova_off = iova_offset(iovad, orig_s_phys); -+ -+ sg_set_page(sgt->sgl, -+ pfn_to_page(PHYS_PFN(orig_s_phys - s_iova_off)), -+ iova_align(iovad, orig_s_len + s_iova_off), -+ sgt->sgl->offset & ~s_iova_off); -+ } else { -+ /* -+ * convince iommu_map_sg_atomic to map the last block -+ * even though it may be too small. -+ */ -+ orig_s_len = last_sg->length; -+ last_sg->length = iova_align(iovad, last_sg->length); -+ } -+ } -+ - ret = iommu_map_sg_atomic(domain, iova, sgt->sgl, sgt->orig_nents, ioprot); -- if (ret < 0 || ret < size) -+ if (ret < 0 || ret < iova_size) - goto out_free_sg; - -+ if (iovad->granule > PAGE_SIZE) { -+ if (size < iovad->granule) { -+ sg_set_page(sgt->sgl, -+ pfn_to_page(PHYS_PFN(orig_s_phys)), -+ orig_s_len, orig_s_off); -+ -+ iova += s_iova_off; -+ } else { -+ last_sg->length = orig_s_len; -+ } -+ } -+ - sgt->sgl->dma_address = iova; -- sgt->sgl->dma_length = size; -+ sgt->sgl->dma_length = iova_size; - return pages; - - out_free_sg: - sg_free_table(sgt); - out_free_iova: -- iommu_dma_free_iova(cookie, iova, size, NULL); -+ iommu_dma_free_iova(cookie, iova, iova_size, NULL); - out_free_pages: - __iommu_dma_free_pages(pages, count); - return NULL; -@@ -1040,8 +1124,9 @@ static int __finalise_sg(struct device *dev, struct scatterlist *sg, int nents, - unsigned int s_length = sg_dma_len(s); - unsigned int s_iova_len = s->length; - -- s->offset += s_iova_off; -- s->length = s_length; -+ sg_set_page(s, -+ pfn_to_page(PHYS_PFN(sg_phys(s) + s_iova_off)), -+ s_length, s_iova_off & ~PAGE_MASK); - sg_dma_address(s) = DMA_MAPPING_ERROR; - sg_dma_len(s) = 0; - -@@ -1082,13 +1167,17 @@ static int __finalise_sg(struct device *dev, struct scatterlist *sg, int nents, - static void __invalidate_sg(struct scatterlist *sg, int nents) - { - struct scatterlist *s; -+ phys_addr_t orig_paddr; - int i; - - for_each_sg(sg, s, nents, i) { -- if (sg_dma_address(s) != DMA_MAPPING_ERROR) -- s->offset += sg_dma_address(s); -- if (sg_dma_len(s)) -- s->length = sg_dma_len(s); -+ if (sg_dma_len(s)) { -+ orig_paddr = sg_phys(s) + sg_dma_address(s); -+ sg_set_page(s, -+ pfn_to_page(PHYS_PFN(orig_paddr)), -+ sg_dma_len(s), -+ sg_dma_address(s) & ~PAGE_MASK); -+ } - sg_dma_address(s) = DMA_MAPPING_ERROR; - sg_dma_len(s) = 0; - } -@@ -1166,15 +1255,16 @@ static int iommu_dma_map_sg(struct device *dev, struct scatterlist *sg, - * stashing the unaligned parts in the as-yet-unused DMA fields. - */ - for_each_sg(sg, s, nents, i) { -- size_t s_iova_off = iova_offset(iovad, s->offset); -+ phys_addr_t s_phys = sg_phys(s); -+ size_t s_iova_off = iova_offset(iovad, s_phys); - size_t s_length = s->length; - size_t pad_len = (mask - iova_len + 1) & mask; - - sg_dma_address(s) = s_iova_off; - sg_dma_len(s) = s_length; -- s->offset -= s_iova_off; - s_length = iova_align(iovad, s_length + s_iova_off); -- s->length = s_length; -+ sg_set_page(s, pfn_to_page(PHYS_PFN(s_phys - s_iova_off)), -+ s_length, s->offset & ~s_iova_off); - - /* - * Due to the alignment of our single IOVA allocation, we can -@@ -1412,9 +1502,15 @@ static int iommu_dma_get_sgtable(struct device *dev, struct sg_table *sgt, - void *cpu_addr, dma_addr_t dma_addr, size_t size, - unsigned long attrs) - { -+ struct iommu_domain *domain = iommu_get_dma_domain(dev); -+ struct iommu_dma_cookie *cookie = domain->iova_cookie; -+ struct iova_domain *iovad = &cookie->iovad; - struct page *page; - int ret; - -+ if (iovad->granule > PAGE_SIZE) -+ return -ENXIO; -+ - if (is_vmalloc_addr(cpu_addr)) { - struct page **pages = dma_common_find_pages(cpu_addr); - -diff --git a/drivers/iommu/iommu.c b/drivers/iommu/iommu.c -index f2c45b85b9fc..0c370e486d6e 100644 ---- a/drivers/iommu/iommu.c -+++ b/drivers/iommu/iommu.c -@@ -80,6 +80,8 @@ static struct iommu_domain *__iommu_domain_alloc(struct bus_type *bus, - unsigned type); - static int __iommu_attach_device(struct iommu_domain *domain, - struct device *dev); -+static void __iommu_detach_device(struct iommu_domain *domain, -+ struct device *dev); - static int __iommu_attach_group(struct iommu_domain *domain, - struct iommu_group *group); - static void __iommu_detach_group(struct iommu_domain *domain, -@@ -1976,6 +1978,24 @@ void iommu_domain_free(struct iommu_domain *domain) - } - EXPORT_SYMBOL_GPL(iommu_domain_free); - -+static int iommu_check_page_size(struct iommu_domain *domain, -+ struct device *dev) -+{ -+ bool trusted = !(dev_is_pci(dev) && to_pci_dev(dev)->untrusted); -+ -+ if (!iommu_is_paging_domain(domain)) -+ return 0; -+ if (iommu_is_large_pages_domain(domain) && trusted) -+ return 0; -+ -+ if (!(domain->pgsize_bitmap & (PAGE_SIZE | (PAGE_SIZE - 1)))) { -+ pr_warn("IOMMU pages cannot exactly represent CPU pages.\n"); -+ return -EFAULT; -+ } -+ -+ return 0; -+} -+ - static int __iommu_attach_device(struct iommu_domain *domain, - struct device *dev) - { -@@ -1985,9 +2005,23 @@ static int __iommu_attach_device(struct iommu_domain *domain, - return -ENODEV; - - ret = domain->ops->attach_dev(domain, dev); -- if (!ret) -- trace_attach_device_to_domain(dev); -- return ret; -+ if (ret) -+ return ret; -+ -+ /* -+ * Check that CPU pages can be represented by the IOVA granularity. -+ * This has to be done after ops->attach_dev since many IOMMU drivers -+ * only limit domain->pgsize_bitmap after having attached the first -+ * device. -+ */ -+ ret = iommu_check_page_size(domain, dev); -+ if (ret) { -+ __iommu_detach_device(domain, dev); -+ return ret; -+ } -+ -+ trace_attach_device_to_domain(dev); -+ return 0; - } - - int iommu_attach_device(struct iommu_domain *domain, struct device *dev) -diff --git a/drivers/iommu/iova.c b/drivers/iommu/iova.c -index db77aa675145..180ce65a6789 100644 ---- a/drivers/iommu/iova.c -+++ b/drivers/iommu/iova.c -@@ -49,10 +49,11 @@ init_iova_domain(struct iova_domain *iovad, unsigned long granule, - { - /* - * IOVA granularity will normally be equal to the smallest -- * supported IOMMU page size; both *must* be capable of -- * representing individual CPU pages exactly. -+ * supported IOMMU page size; while both usually are capable of -+ * representing individual CPU pages exactly the IOVA allocator -+ * supports any granularities that are an exact power of two. - */ -- BUG_ON((granule > PAGE_SIZE) || !is_power_of_2(granule)); -+ BUG_ON(!is_power_of_2(granule)); - - spin_lock_init(&iovad->iova_rbtree_lock); - iovad->rbroot = RB_ROOT; -diff --git a/include/linux/iommu.h b/include/linux/iommu.h -index 9208eca4b0d1..dec2dd70a876 100644 ---- a/include/linux/iommu.h -+++ b/include/linux/iommu.h -@@ -63,6 +63,8 @@ struct iommu_domain_geometry { - implementation */ - #define __IOMMU_DOMAIN_PT (1U << 2) /* Domain is identity mapped */ - #define __IOMMU_DOMAIN_DMA_FQ (1U << 3) /* DMA-API uses flush queue */ -+#define __IOMMU_DOMAIN_LP (1U << 4) /* Support for PAGE_SIZE smaller -+ than IOMMU page size */ - - /* - * This are the possible domain-types -@@ -82,10 +84,12 @@ struct iommu_domain_geometry { - #define IOMMU_DOMAIN_IDENTITY (__IOMMU_DOMAIN_PT) - #define IOMMU_DOMAIN_UNMANAGED (__IOMMU_DOMAIN_PAGING) - #define IOMMU_DOMAIN_DMA (__IOMMU_DOMAIN_PAGING | \ -- __IOMMU_DOMAIN_DMA_API) -+ __IOMMU_DOMAIN_DMA_API | \ -+ __IOMMU_DOMAIN_LP) - #define IOMMU_DOMAIN_DMA_FQ (__IOMMU_DOMAIN_PAGING | \ - __IOMMU_DOMAIN_DMA_API | \ -- __IOMMU_DOMAIN_DMA_FQ) -+ __IOMMU_DOMAIN_DMA_FQ | \ -+ __IOMMU_DOMAIN_LP) - - struct iommu_domain { - unsigned type; -@@ -102,6 +106,16 @@ static inline bool iommu_is_dma_domain(struct iommu_domain *domain) - return domain->type & __IOMMU_DOMAIN_DMA_API; - } - -+static inline bool iommu_is_paging_domain(struct iommu_domain *domain) -+{ -+ return domain->type & __IOMMU_DOMAIN_PAGING; -+} -+ -+static inline bool iommu_is_large_pages_domain(struct iommu_domain *domain) -+{ -+ return domain->type & __IOMMU_DOMAIN_LP; -+} -+ - enum iommu_cap { - IOMMU_CAP_CACHE_COHERENCY, /* IOMMU can enforce cache coherent DMA - transactions */ -- cgit 1.4.1