@@ -1538,13 +1538,13 @@ static int suspend_vq(struct mlx5_vdpa_net *ndev, struct mlx5_vdpa_virtqueue *mv
1538
1538
1539
1539
err = modify_virtqueue_state (ndev , mvq , MLX5_VIRTIO_NET_Q_OBJECT_STATE_SUSPEND );
1540
1540
if (err ) {
1541
- mlx5_vdpa_warn (& ndev -> mvdev , "modify to suspend failed, err: %d\n" , err );
1541
+ mlx5_vdpa_err (& ndev -> mvdev , "modify to suspend failed, err: %d\n" , err );
1542
1542
return err ;
1543
1543
}
1544
1544
1545
1545
err = query_virtqueue (ndev , mvq , & attr );
1546
1546
if (err ) {
1547
- mlx5_vdpa_warn (& ndev -> mvdev , "failed to query virtqueue, err: %d\n" , err );
1547
+ mlx5_vdpa_err (& ndev -> mvdev , "failed to query virtqueue, err: %d\n" , err );
1548
1548
return err ;
1549
1549
}
1550
1550
@@ -1585,7 +1585,7 @@ static int resume_vq(struct mlx5_vdpa_net *ndev, struct mlx5_vdpa_virtqueue *mvq
1585
1585
*/
1586
1586
err = modify_virtqueue (ndev , mvq , 0 );
1587
1587
if (err ) {
1588
- mlx5_vdpa_warn (& ndev -> mvdev ,
1588
+ mlx5_vdpa_err (& ndev -> mvdev ,
1589
1589
"modify vq properties failed for vq %u, err: %d\n" ,
1590
1590
mvq -> index , err );
1591
1591
return err ;
@@ -1600,15 +1600,15 @@ static int resume_vq(struct mlx5_vdpa_net *ndev, struct mlx5_vdpa_virtqueue *mvq
1600
1600
case MLX5_VIRTIO_NET_Q_OBJECT_STATE_RDY :
1601
1601
return 0 ;
1602
1602
default :
1603
- mlx5_vdpa_warn (& ndev -> mvdev , "resume vq %u called from bad state %d\n" ,
1603
+ mlx5_vdpa_err (& ndev -> mvdev , "resume vq %u called from bad state %d\n" ,
1604
1604
mvq -> index , mvq -> fw_state );
1605
1605
return - EINVAL ;
1606
1606
}
1607
1607
1608
1608
err = modify_virtqueue_state (ndev , mvq , MLX5_VIRTIO_NET_Q_OBJECT_STATE_RDY );
1609
1609
if (err )
1610
- mlx5_vdpa_warn (& ndev -> mvdev , "modify to resume failed for vq %u, err: %d\n" ,
1611
- mvq -> index , err );
1610
+ mlx5_vdpa_err (& ndev -> mvdev , "modify to resume failed for vq %u, err: %d\n" ,
1611
+ mvq -> index , err );
1612
1612
1613
1613
return err ;
1614
1614
}
@@ -2002,13 +2002,13 @@ static int setup_steering(struct mlx5_vdpa_net *ndev)
2002
2002
2003
2003
ns = mlx5_get_flow_namespace (ndev -> mvdev .mdev , MLX5_FLOW_NAMESPACE_BYPASS );
2004
2004
if (!ns ) {
2005
- mlx5_vdpa_warn (& ndev -> mvdev , "failed to get flow namespace\n" );
2005
+ mlx5_vdpa_err (& ndev -> mvdev , "failed to get flow namespace\n" );
2006
2006
return - EOPNOTSUPP ;
2007
2007
}
2008
2008
2009
2009
ndev -> rxft = mlx5_create_auto_grouped_flow_table (ns , & ft_attr );
2010
2010
if (IS_ERR (ndev -> rxft )) {
2011
- mlx5_vdpa_warn (& ndev -> mvdev , "failed to create flow table\n" );
2011
+ mlx5_vdpa_err (& ndev -> mvdev , "failed to create flow table\n" );
2012
2012
return PTR_ERR (ndev -> rxft );
2013
2013
}
2014
2014
mlx5_vdpa_add_rx_flow_table (ndev );
@@ -2530,7 +2530,7 @@ static int mlx5_vdpa_get_vq_state(struct vdpa_device *vdev, u16 idx, struct vdpa
2530
2530
2531
2531
err = query_virtqueue (ndev , mvq , & attr );
2532
2532
if (err ) {
2533
- mlx5_vdpa_warn (mvdev , "failed to query virtqueue\n" );
2533
+ mlx5_vdpa_err (mvdev , "failed to query virtqueue\n" );
2534
2534
return err ;
2535
2535
}
2536
2536
state -> split .avail_index = attr .used_index ;
@@ -3189,7 +3189,7 @@ static int mlx5_vdpa_compat_reset(struct vdpa_device *vdev, u32 flags)
3189
3189
if ((flags & VDPA_RESET_F_CLEAN_MAP ) &&
3190
3190
MLX5_CAP_GEN (mvdev -> mdev , umem_uid_0 )) {
3191
3191
if (mlx5_vdpa_create_dma_mr (mvdev ))
3192
- mlx5_vdpa_warn (mvdev , "create MR failed\n" );
3192
+ mlx5_vdpa_err (mvdev , "create MR failed\n" );
3193
3193
}
3194
3194
if (vq_reset )
3195
3195
setup_vq_resources (ndev , false);
@@ -3244,7 +3244,7 @@ static int set_map_data(struct mlx5_vdpa_dev *mvdev, struct vhost_iotlb *iotlb,
3244
3244
new_mr = mlx5_vdpa_create_mr (mvdev , iotlb );
3245
3245
if (IS_ERR (new_mr )) {
3246
3246
err = PTR_ERR (new_mr );
3247
- mlx5_vdpa_warn (mvdev , "create map failed(%d)\n" , err );
3247
+ mlx5_vdpa_err (mvdev , "create map failed(%d)\n" , err );
3248
3248
return err ;
3249
3249
}
3250
3250
} else {
@@ -3257,7 +3257,7 @@ static int set_map_data(struct mlx5_vdpa_dev *mvdev, struct vhost_iotlb *iotlb,
3257
3257
} else {
3258
3258
err = mlx5_vdpa_change_map (mvdev , new_mr , asid );
3259
3259
if (err ) {
3260
- mlx5_vdpa_warn (mvdev , "change map failed(%d)\n" , err );
3260
+ mlx5_vdpa_err (mvdev , "change map failed(%d)\n" , err );
3261
3261
goto out_err ;
3262
3262
}
3263
3263
}
0 commit comments