@@ -80,7 +80,7 @@ vmalloc_done_load:
80
80
* see if we need to jump to huge tlb processing.
81
81
*/
82
82
andi t0, ra, _PAGE_HUGE
83
- bne t0 , zero , tlb_huge_update_load
83
+ bnez t0, tlb_huge_update_load
84
84
85
85
csrrd t0, LOONGARCH_CSR_BADV
86
86
srli.d t0, t0, (PAGE_SHIFT + PTE_ORDER)
@@ -100,12 +100,12 @@ smp_pgtable_change_load:
100
100
101
101
srli.d ra, t0, _PAGE_PRESENT_SHIFT
102
102
andi ra, ra, 1
103
- beq ra , zero , nopage_tlb_load
103
+ beqz ra, nopage_tlb_load
104
104
105
105
ori t0, t0, _PAGE_VALID
106
106
#ifdef CONFIG_SMP
107
107
sc.d t0, t1, 0
108
- beq t0 , zero , smp_pgtable_change_load
108
+ beqz t0, smp_pgtable_change_load
109
109
#else
110
110
st.d t0, t1, 0
111
111
#endif
@@ -139,13 +139,13 @@ tlb_huge_update_load:
139
139
#endif
140
140
srli.d ra, t0, _PAGE_PRESENT_SHIFT
141
141
andi ra, ra, 1
142
- beq ra , zero , nopage_tlb_load
142
+ beqz ra, nopage_tlb_load
143
143
tlbsrch
144
144
145
145
ori t0, t0, _PAGE_VALID
146
146
#ifdef CONFIG_SMP
147
147
sc.d t0, t1, 0
148
- beq t0 , zero , tlb_huge_update_load
148
+ beqz t0, tlb_huge_update_load
149
149
ld.d t0, t1, 0
150
150
#else
151
151
st.d t0, t1, 0
@@ -244,7 +244,7 @@ vmalloc_done_store:
244
244
* see if we need to jump to huge tlb processing.
245
245
*/
246
246
andi t0, ra, _PAGE_HUGE
247
- bne t0 , zero , tlb_huge_update_store
247
+ bnez t0, tlb_huge_update_store
248
248
249
249
csrrd t0, LOONGARCH_CSR_BADV
250
250
srli.d t0, t0, (PAGE_SHIFT + PTE_ORDER)
@@ -265,12 +265,12 @@ smp_pgtable_change_store:
265
265
srli.d ra, t0, _PAGE_PRESENT_SHIFT
266
266
andi ra, ra, ((_PAGE_PRESENT | _PAGE_WRITE) >> _PAGE_PRESENT_SHIFT)
267
267
xori ra, ra, ((_PAGE_PRESENT | _PAGE_WRITE) >> _PAGE_PRESENT_SHIFT)
268
- bne ra , zero , nopage_tlb_store
268
+ bnez ra, nopage_tlb_store
269
269
270
270
ori t0, t0, (_PAGE_VALID | _PAGE_DIRTY | _PAGE_MODIFIED)
271
271
#ifdef CONFIG_SMP
272
272
sc.d t0, t1, 0
273
- beq t0 , zero , smp_pgtable_change_store
273
+ beqz t0, smp_pgtable_change_store
274
274
#else
275
275
st.d t0, t1, 0
276
276
#endif
@@ -306,14 +306,14 @@ tlb_huge_update_store:
306
306
srli.d ra, t0, _PAGE_PRESENT_SHIFT
307
307
andi ra, ra, ((_PAGE_PRESENT | _PAGE_WRITE) >> _PAGE_PRESENT_SHIFT)
308
308
xori ra, ra, ((_PAGE_PRESENT | _PAGE_WRITE) >> _PAGE_PRESENT_SHIFT)
309
- bne ra , zero , nopage_tlb_store
309
+ bnez ra, nopage_tlb_store
310
310
311
311
tlbsrch
312
312
ori t0, t0, (_PAGE_VALID | _PAGE_DIRTY | _PAGE_MODIFIED)
313
313
314
314
#ifdef CONFIG_SMP
315
315
sc.d t0, t1, 0
316
- beq t0 , zero , tlb_huge_update_store
316
+ beqz t0, tlb_huge_update_store
317
317
ld.d t0, t1, 0
318
318
#else
319
319
st.d t0, t1, 0
@@ -411,7 +411,7 @@ vmalloc_done_modify:
411
411
* see if we need to jump to huge tlb processing.
412
412
*/
413
413
andi t0, ra, _PAGE_HUGE
414
- bne t0 , zero , tlb_huge_update_modify
414
+ bnez t0, tlb_huge_update_modify
415
415
416
416
csrrd t0, LOONGARCH_CSR_BADV
417
417
srli.d t0, t0, (PAGE_SHIFT + PTE_ORDER)
@@ -431,12 +431,12 @@ smp_pgtable_change_modify:
431
431
432
432
srli.d ra, t0, _PAGE_WRITE_SHIFT
433
433
andi ra, ra, 1
434
- beq ra , zero , nopage_tlb_modify
434
+ beqz ra, nopage_tlb_modify
435
435
436
436
ori t0, t0, (_PAGE_VALID | _PAGE_DIRTY | _PAGE_MODIFIED)
437
437
#ifdef CONFIG_SMP
438
438
sc.d t0, t1, 0
439
- beq t0 , zero , smp_pgtable_change_modify
439
+ beqz t0, smp_pgtable_change_modify
440
440
#else
441
441
st.d t0, t1, 0
442
442
#endif
@@ -471,14 +471,14 @@ tlb_huge_update_modify:
471
471
472
472
srli.d ra, t0, _PAGE_WRITE_SHIFT
473
473
andi ra, ra, 1
474
- beq ra , zero , nopage_tlb_modify
474
+ beqz ra, nopage_tlb_modify
475
475
476
476
tlbsrch
477
477
ori t0, t0, (_PAGE_VALID | _PAGE_DIRTY | _PAGE_MODIFIED)
478
478
479
479
#ifdef CONFIG_SMP
480
480
sc.d t0, t1, 0
481
- beq t0 , zero , tlb_huge_update_modify
481
+ beqz t0, tlb_huge_update_modify
482
482
ld.d t0, t1, 0
483
483
#else
484
484
st.d t0, t1, 0
0 commit comments