@@ -351,7 +351,7 @@ static int blk_validate_limits(struct queue_limits *lim)
351
351
352
352
if (lim -> alignment_offset ) {
353
353
lim -> alignment_offset &= (lim -> physical_block_size - 1 );
354
- lim -> features &= ~BLK_FEAT_MISALIGNED ;
354
+ lim -> flags &= ~BLK_FLAG_MISALIGNED ;
355
355
}
356
356
357
357
if (!(lim -> features & BLK_FEAT_WRITE_CACHE ))
@@ -564,7 +564,7 @@ int blk_stack_limits(struct queue_limits *t, struct queue_limits *b,
564
564
if (!(b -> features & BLK_FEAT_POLL ))
565
565
t -> features &= ~BLK_FEAT_POLL ;
566
566
567
- t -> flags |= (b -> flags & BLK_FEAT_MISALIGNED );
567
+ t -> flags |= (b -> flags & BLK_FLAG_MISALIGNED );
568
568
569
569
t -> max_sectors = min_not_zero (t -> max_sectors , b -> max_sectors );
570
570
t -> max_user_sectors = min_not_zero (t -> max_user_sectors ,
@@ -603,7 +603,7 @@ int blk_stack_limits(struct queue_limits *t, struct queue_limits *b,
603
603
604
604
/* Verify that top and bottom intervals line up */
605
605
if (max (top , bottom ) % min (top , bottom )) {
606
- t -> flags |= BLK_FEAT_MISALIGNED ;
606
+ t -> flags |= BLK_FLAG_MISALIGNED ;
607
607
ret = -1 ;
608
608
}
609
609
}
@@ -625,28 +625,28 @@ int blk_stack_limits(struct queue_limits *t, struct queue_limits *b,
625
625
/* Physical block size a multiple of the logical block size? */
626
626
if (t -> physical_block_size & (t -> logical_block_size - 1 )) {
627
627
t -> physical_block_size = t -> logical_block_size ;
628
- t -> flags |= BLK_FEAT_MISALIGNED ;
628
+ t -> flags |= BLK_FLAG_MISALIGNED ;
629
629
ret = -1 ;
630
630
}
631
631
632
632
/* Minimum I/O a multiple of the physical block size? */
633
633
if (t -> io_min & (t -> physical_block_size - 1 )) {
634
634
t -> io_min = t -> physical_block_size ;
635
- t -> flags |= BLK_FEAT_MISALIGNED ;
635
+ t -> flags |= BLK_FLAG_MISALIGNED ;
636
636
ret = -1 ;
637
637
}
638
638
639
639
/* Optimal I/O a multiple of the physical block size? */
640
640
if (t -> io_opt & (t -> physical_block_size - 1 )) {
641
641
t -> io_opt = 0 ;
642
- t -> flags |= BLK_FEAT_MISALIGNED ;
642
+ t -> flags |= BLK_FLAG_MISALIGNED ;
643
643
ret = -1 ;
644
644
}
645
645
646
646
/* chunk_sectors a multiple of the physical block size? */
647
647
if ((t -> chunk_sectors << 9 ) & (t -> physical_block_size - 1 )) {
648
648
t -> chunk_sectors = 0 ;
649
- t -> flags |= BLK_FEAT_MISALIGNED ;
649
+ t -> flags |= BLK_FLAG_MISALIGNED ;
650
650
ret = -1 ;
651
651
}
652
652
@@ -656,7 +656,7 @@ int blk_stack_limits(struct queue_limits *t, struct queue_limits *b,
656
656
657
657
/* Verify that new alignment_offset is on a logical block boundary */
658
658
if (t -> alignment_offset & (t -> logical_block_size - 1 )) {
659
- t -> flags |= BLK_FEAT_MISALIGNED ;
659
+ t -> flags |= BLK_FLAG_MISALIGNED ;
660
660
ret = -1 ;
661
661
}
662
662
@@ -809,7 +809,7 @@ int bdev_alignment_offset(struct block_device *bdev)
809
809
{
810
810
struct request_queue * q = bdev_get_queue (bdev );
811
811
812
- if (q -> limits .flags & BLK_FEAT_MISALIGNED )
812
+ if (q -> limits .flags & BLK_FLAG_MISALIGNED )
813
813
return -1 ;
814
814
if (bdev_is_partition (bdev ))
815
815
return queue_limit_alignment_offset (& q -> limits ,
0 commit comments