@@ -327,44 +327,42 @@ static void mtk_iommu_v1_detach_device(struct iommu_domain *domain, struct devic
327
327
}
328
328
329
329
static int mtk_iommu_v1_map (struct iommu_domain * domain , unsigned long iova ,
330
- phys_addr_t paddr , size_t size , int prot , gfp_t gfp )
330
+ phys_addr_t paddr , size_t pgsize , size_t pgcount ,
331
+ int prot , gfp_t gfp , size_t * mapped )
331
332
{
332
333
struct mtk_iommu_v1_domain * dom = to_mtk_domain (domain );
333
- unsigned int page_num = size >> MT2701_IOMMU_PAGE_SHIFT ;
334
334
unsigned long flags ;
335
335
unsigned int i ;
336
336
u32 * pgt_base_iova = dom -> pgt_va + (iova >> MT2701_IOMMU_PAGE_SHIFT );
337
337
u32 pabase = (u32 )paddr ;
338
- int map_size = 0 ;
339
338
340
339
spin_lock_irqsave (& dom -> pgtlock , flags );
341
- for (i = 0 ; i < page_num ; i ++ ) {
342
- if (pgt_base_iova [i ]) {
343
- memset (pgt_base_iova , 0 , i * sizeof (u32 ));
340
+ for (i = 0 ; i < pgcount ; i ++ ) {
341
+ if (pgt_base_iova [i ])
344
342
break ;
345
- }
346
343
pgt_base_iova [i ] = pabase | F_DESC_VALID | F_DESC_NONSEC ;
347
344
pabase += MT2701_IOMMU_PAGE_SIZE ;
348
- map_size += MT2701_IOMMU_PAGE_SIZE ;
349
345
}
350
346
351
347
spin_unlock_irqrestore (& dom -> pgtlock , flags );
352
348
353
- mtk_iommu_v1_tlb_flush_range (dom -> data , iova , size );
349
+ * mapped = i * MT2701_IOMMU_PAGE_SIZE ;
350
+ mtk_iommu_v1_tlb_flush_range (dom -> data , iova , * mapped );
354
351
355
- return map_size == size ? 0 : - EEXIST ;
352
+ return i == pgcount ? 0 : - EEXIST ;
356
353
}
357
354
358
355
static size_t mtk_iommu_v1_unmap (struct iommu_domain * domain , unsigned long iova ,
359
- size_t size , struct iommu_iotlb_gather * gather )
356
+ size_t pgsize , size_t pgcount ,
357
+ struct iommu_iotlb_gather * gather )
360
358
{
361
359
struct mtk_iommu_v1_domain * dom = to_mtk_domain (domain );
362
360
unsigned long flags ;
363
361
u32 * pgt_base_iova = dom -> pgt_va + (iova >> MT2701_IOMMU_PAGE_SHIFT );
364
- unsigned int page_num = size >> MT2701_IOMMU_PAGE_SHIFT ;
362
+ size_t size = pgcount * MT2701_IOMMU_PAGE_SIZE ;
365
363
366
364
spin_lock_irqsave (& dom -> pgtlock , flags );
367
- memset (pgt_base_iova , 0 , page_num * sizeof (u32 ));
365
+ memset (pgt_base_iova , 0 , pgcount * sizeof (u32 ));
368
366
spin_unlock_irqrestore (& dom -> pgtlock , flags );
369
367
370
368
mtk_iommu_v1_tlb_flush_range (dom -> data , iova , size );
@@ -586,13 +584,13 @@ static const struct iommu_ops mtk_iommu_v1_ops = {
586
584
.release_device = mtk_iommu_v1_release_device ,
587
585
.def_domain_type = mtk_iommu_v1_def_domain_type ,
588
586
.device_group = generic_device_group ,
589
- .pgsize_bitmap = ~ 0UL << MT2701_IOMMU_PAGE_SHIFT ,
587
+ .pgsize_bitmap = MT2701_IOMMU_PAGE_SIZE ,
590
588
.owner = THIS_MODULE ,
591
589
.default_domain_ops = & (const struct iommu_domain_ops ) {
592
590
.attach_dev = mtk_iommu_v1_attach_device ,
593
591
.detach_dev = mtk_iommu_v1_detach_device ,
594
- .map = mtk_iommu_v1_map ,
595
- .unmap = mtk_iommu_v1_unmap ,
592
+ .map_pages = mtk_iommu_v1_map ,
593
+ .unmap_pages = mtk_iommu_v1_unmap ,
596
594
.iova_to_phys = mtk_iommu_v1_iova_to_phys ,
597
595
.free = mtk_iommu_v1_domain_free ,
598
596
}
0 commit comments