@@ -214,7 +214,7 @@ static void __init clear_pgds(unsigned long start,
214
214
static void __init kasan_init_shadow (void )
215
215
{
216
216
u64 kimg_shadow_start , kimg_shadow_end ;
217
- u64 mod_shadow_start , mod_shadow_end ;
217
+ u64 mod_shadow_start ;
218
218
u64 vmalloc_shadow_end ;
219
219
phys_addr_t pa_start , pa_end ;
220
220
u64 i ;
@@ -223,7 +223,6 @@ static void __init kasan_init_shadow(void)
223
223
kimg_shadow_end = PAGE_ALIGN ((u64 )kasan_mem_to_shadow (KERNEL_END ));
224
224
225
225
mod_shadow_start = (u64 )kasan_mem_to_shadow ((void * )MODULES_VADDR );
226
- mod_shadow_end = (u64 )kasan_mem_to_shadow ((void * )MODULES_END );
227
226
228
227
vmalloc_shadow_end = (u64 )kasan_mem_to_shadow ((void * )VMALLOC_END );
229
228
@@ -246,17 +245,9 @@ static void __init kasan_init_shadow(void)
246
245
kasan_populate_early_shadow (kasan_mem_to_shadow ((void * )PAGE_END ),
247
246
(void * )mod_shadow_start );
248
247
249
- if (IS_ENABLED (CONFIG_KASAN_VMALLOC )) {
250
- BUILD_BUG_ON (VMALLOC_START != MODULES_END );
251
- kasan_populate_early_shadow ((void * )vmalloc_shadow_end ,
252
- (void * )KASAN_SHADOW_END );
253
- } else {
254
- kasan_populate_early_shadow ((void * )kimg_shadow_end ,
255
- (void * )KASAN_SHADOW_END );
256
- if (kimg_shadow_start > mod_shadow_end )
257
- kasan_populate_early_shadow ((void * )mod_shadow_end ,
258
- (void * )kimg_shadow_start );
259
- }
248
+ BUILD_BUG_ON (VMALLOC_START != MODULES_END );
249
+ kasan_populate_early_shadow ((void * )vmalloc_shadow_end ,
250
+ (void * )KASAN_SHADOW_END );
260
251
261
252
for_each_mem_range (i , & pa_start , & pa_end ) {
262
253
void * start = (void * )__phys_to_virt (pa_start );
0 commit comments