@@ -273,13 +273,13 @@ int map_kernel_range_noflush(unsigned long addr, unsigned long size,
273
273
return nr ;
274
274
}
275
275
276
- static int vmap_page_range (unsigned long start , unsigned long end ,
276
+ static int map_kernel_range (unsigned long start , unsigned long size ,
277
277
pgprot_t prot , struct page * * pages )
278
278
{
279
279
int ret ;
280
280
281
- ret = map_kernel_range_noflush (start , end - start , prot , pages );
282
- flush_cache_vmap (start , end );
281
+ ret = map_kernel_range_noflush (start , size , prot , pages );
282
+ flush_cache_vmap (start , start + size );
283
283
return ret ;
284
284
}
285
285
@@ -1867,7 +1867,7 @@ void *vm_map_ram(struct page **pages, unsigned int count, int node, pgprot_t pro
1867
1867
1868
1868
kasan_unpoison_vmalloc (mem , size );
1869
1869
1870
- if (vmap_page_range (addr , addr + size , prot , pages ) < 0 ) {
1870
+ if (map_kernel_range (addr , size , prot , pages ) < 0 ) {
1871
1871
vm_unmap_ram (mem , count );
1872
1872
return NULL ;
1873
1873
}
@@ -2031,10 +2031,9 @@ void unmap_kernel_range(unsigned long addr, unsigned long size)
2031
2031
int map_vm_area (struct vm_struct * area , pgprot_t prot , struct page * * pages )
2032
2032
{
2033
2033
unsigned long addr = (unsigned long )area -> addr ;
2034
- unsigned long end = addr + get_vm_area_size (area );
2035
2034
int err ;
2036
2035
2037
- err = vmap_page_range (addr , end , prot , pages );
2036
+ err = map_kernel_range (addr , get_vm_area_size ( area ) , prot , pages );
2038
2037
2039
2038
return err > 0 ? 0 : err ;
2040
2039
}
0 commit comments