@@ -306,8 +306,6 @@ static void iomap_finish_folio_read(struct folio *folio, size_t off,
306
306
spin_unlock_irqrestore (& ifs -> state_lock , flags );
307
307
}
308
308
309
- if (error )
310
- folio_set_error (folio );
311
309
if (finished )
312
310
folio_end_read (folio , uptodate );
313
311
}
@@ -460,9 +458,6 @@ int iomap_read_folio(struct folio *folio, const struct iomap_ops *ops)
460
458
while ((ret = iomap_iter (& iter , ops )) > 0 )
461
459
iter .processed = iomap_readpage_iter (& iter , & ctx , 0 );
462
460
463
- if (ret < 0 )
464
- folio_set_error (folio );
465
-
466
461
if (ctx .bio ) {
467
462
submit_bio (ctx .bio );
468
463
WARN_ON_ONCE (!ctx .cur_folio_in_bio );
@@ -697,7 +692,6 @@ static int __iomap_write_begin(const struct iomap_iter *iter, loff_t pos,
697
692
698
693
if (folio_test_uptodate (folio ))
699
694
return 0 ;
700
- folio_clear_error (folio );
701
695
702
696
do {
703
697
iomap_adjust_read_range (iter -> inode , folio , & block_start ,
@@ -1543,8 +1537,6 @@ iomap_finish_ioend(struct iomap_ioend *ioend, int error)
1543
1537
1544
1538
/* walk all folios in bio, ending page IO on them */
1545
1539
bio_for_each_folio_all (fi , bio ) {
1546
- if (error )
1547
- folio_set_error (fi .folio );
1548
1540
iomap_finish_folio_write (inode , fi .folio , fi .length );
1549
1541
folio_count ++ ;
1550
1542
}
0 commit comments