@@ -1191,7 +1191,7 @@ void __ceph_allocate_page_array(struct ceph_writeback_ctl *ceph_wbc,
1191
1191
static inline
1192
1192
void ceph_allocate_page_array (struct address_space * mapping ,
1193
1193
struct ceph_writeback_ctl * ceph_wbc ,
1194
- struct page * page )
1194
+ struct folio * folio )
1195
1195
{
1196
1196
struct inode * inode = mapping -> host ;
1197
1197
struct ceph_inode_info * ci = ceph_inode (inode );
@@ -1200,13 +1200,13 @@ void ceph_allocate_page_array(struct address_space *mapping,
1200
1200
u32 xlen ;
1201
1201
1202
1202
/* prepare async write request */
1203
- ceph_wbc -> offset = (u64 )page_offset ( page );
1203
+ ceph_wbc -> offset = (u64 )folio_pos ( folio );
1204
1204
ceph_calc_file_object_mapping (& ci -> i_layout ,
1205
1205
ceph_wbc -> offset , ceph_wbc -> wsize ,
1206
1206
& objnum , & objoff , & xlen );
1207
1207
1208
1208
ceph_wbc -> num_ops = 1 ;
1209
- ceph_wbc -> strip_unit_end = page -> index + ((xlen - 1 ) >> PAGE_SHIFT );
1209
+ ceph_wbc -> strip_unit_end = folio -> index + ((xlen - 1 ) >> PAGE_SHIFT );
1210
1210
1211
1211
BUG_ON (ceph_wbc -> pages );
1212
1212
ceph_wbc -> max_pages = calc_pages_for (0 , (u64 )xlen );
@@ -1338,7 +1338,7 @@ int ceph_process_folio_batch(struct address_space *mapping,
1338
1338
* allocate a page array
1339
1339
*/
1340
1340
if (ceph_wbc -> locked_pages == 0 ) {
1341
- ceph_allocate_page_array (mapping , ceph_wbc , & folio -> page );
1341
+ ceph_allocate_page_array (mapping , ceph_wbc , folio );
1342
1342
} else if (!is_folio_index_contiguous (ceph_wbc , folio )) {
1343
1343
if (is_num_ops_too_big (ceph_wbc )) {
1344
1344
folio_redirty_for_writepage (wbc , folio );
0 commit comments