@@ -373,7 +373,7 @@ static int upgrade_mode(struct dm_dev_internal *dd, fmode_t new_mode,
373
373
*/
374
374
dev_t dm_get_dev_t (const char * path )
375
375
{
376
- dev_t uninitialized_var ( dev ) ;
376
+ dev_t dev ;
377
377
struct block_device * bdev ;
378
378
379
379
bdev = lookup_bdev (path );
@@ -627,13 +627,13 @@ static int validate_hardware_logical_block_alignment(struct dm_table *table,
627
627
628
628
struct dm_target * uninitialized_var (ti );
629
629
struct queue_limits ti_limits ;
630
- unsigned i = 0 ;
630
+ unsigned i ;
631
631
632
632
/*
633
633
* Check each entry in the table in turn.
634
634
*/
635
- while (i < dm_table_get_num_targets (table )) {
636
- ti = dm_table_get_target (table , i ++ );
635
+ for (i = 0 ; i < dm_table_get_num_targets (table ); i ++ ) {
636
+ ti = dm_table_get_target (table , i );
637
637
638
638
blk_set_stacking_limits (& ti_limits );
639
639
@@ -854,11 +854,11 @@ static int device_supports_dax(struct dm_target *ti, struct dm_dev *dev,
854
854
static bool dm_table_supports_dax (struct dm_table * t )
855
855
{
856
856
struct dm_target * ti ;
857
- unsigned i = 0 ;
857
+ unsigned i ;
858
858
859
859
/* Ensure that all targets support DAX. */
860
- while (i < dm_table_get_num_targets (t )) {
861
- ti = dm_table_get_target (t , i ++ );
860
+ for (i = 0 ; i < dm_table_get_num_targets (t ); i ++ ) {
861
+ ti = dm_table_get_target (t , i );
862
862
863
863
if (!ti -> type -> direct_access )
864
864
return false;
@@ -1010,11 +1010,11 @@ struct dm_target *dm_table_get_immutable_target(struct dm_table *t)
1010
1010
1011
1011
struct dm_target * dm_table_get_wildcard_target (struct dm_table * t )
1012
1012
{
1013
- struct dm_target * uninitialized_var ( ti ) ;
1014
- unsigned i = 0 ;
1013
+ struct dm_target * ti ;
1014
+ unsigned i ;
1015
1015
1016
- while (i < dm_table_get_num_targets (t )) {
1017
- ti = dm_table_get_target (t , i ++ );
1016
+ for (i = 0 ; i < dm_table_get_num_targets (t ); i ++ ) {
1017
+ ti = dm_table_get_target (t , i );
1018
1018
if (dm_target_is_wildcard (ti -> type ))
1019
1019
return ti ;
1020
1020
}
@@ -1321,15 +1321,16 @@ static int count_device(struct dm_target *ti, struct dm_dev *dev,
1321
1321
*/
1322
1322
bool dm_table_has_no_data_devices (struct dm_table * table )
1323
1323
{
1324
- struct dm_target * uninitialized_var ( ti ) ;
1325
- unsigned i = 0 , num_devices = 0 ;
1324
+ struct dm_target * ti ;
1325
+ unsigned i , num_devices ;
1326
1326
1327
- while (i < dm_table_get_num_targets (table )) {
1328
- ti = dm_table_get_target (table , i ++ );
1327
+ for (i = 0 ; i < dm_table_get_num_targets (table ); i ++ ) {
1328
+ ti = dm_table_get_target (table , i );
1329
1329
1330
1330
if (!ti -> type -> iterate_devices )
1331
1331
return false;
1332
1332
1333
+ num_devices = 0 ;
1333
1334
ti -> type -> iterate_devices (ti , count_device , & num_devices );
1334
1335
if (num_devices )
1335
1336
return false;
@@ -1344,16 +1345,16 @@ bool dm_table_has_no_data_devices(struct dm_table *table)
1344
1345
int dm_calculate_queue_limits (struct dm_table * table ,
1345
1346
struct queue_limits * limits )
1346
1347
{
1347
- struct dm_target * uninitialized_var ( ti ) ;
1348
+ struct dm_target * ti ;
1348
1349
struct queue_limits ti_limits ;
1349
- unsigned i = 0 ;
1350
+ unsigned i ;
1350
1351
1351
1352
blk_set_stacking_limits (limits );
1352
1353
1353
- while (i < dm_table_get_num_targets (table )) {
1354
+ for (i = 0 ; i < dm_table_get_num_targets (table ); i ++ ) {
1354
1355
blk_set_stacking_limits (& ti_limits );
1355
1356
1356
- ti = dm_table_get_target (table , i ++ );
1357
+ ti = dm_table_get_target (table , i );
1357
1358
1358
1359
if (!ti -> type -> iterate_devices )
1359
1360
goto combine_limits ;
@@ -1435,16 +1436,16 @@ static int device_flush_capable(struct dm_target *ti, struct dm_dev *dev,
1435
1436
static bool dm_table_supports_flush (struct dm_table * t , unsigned long flush )
1436
1437
{
1437
1438
struct dm_target * ti ;
1438
- unsigned i = 0 ;
1439
+ unsigned i ;
1439
1440
1440
1441
/*
1441
1442
* Require at least one underlying device to support flushes.
1442
1443
* t->devices includes internal dm devices such as mirror logs
1443
1444
* so we need to use iterate_devices here, which targets
1444
1445
* supporting flushes must provide.
1445
1446
*/
1446
- while (i < dm_table_get_num_targets (t )) {
1447
- ti = dm_table_get_target (t , i ++ );
1447
+ for (i = 0 ; i < dm_table_get_num_targets (t ); i ++ ) {
1448
+ ti = dm_table_get_target (t , i );
1448
1449
1449
1450
if (!ti -> num_flush_bios )
1450
1451
continue ;
@@ -1504,10 +1505,10 @@ static bool dm_table_all_devices_attribute(struct dm_table *t,
1504
1505
iterate_devices_callout_fn func )
1505
1506
{
1506
1507
struct dm_target * ti ;
1507
- unsigned i = 0 ;
1508
+ unsigned i ;
1508
1509
1509
- while (i < dm_table_get_num_targets (t )) {
1510
- ti = dm_table_get_target (t , i ++ );
1510
+ for (i = 0 ; i < dm_table_get_num_targets (t ); i ++ ) {
1511
+ ti = dm_table_get_target (t , i );
1511
1512
1512
1513
if (!ti -> type -> iterate_devices ||
1513
1514
!ti -> type -> iterate_devices (ti , func , NULL ))
@@ -1528,10 +1529,10 @@ static int device_not_write_same_capable(struct dm_target *ti, struct dm_dev *de
1528
1529
static bool dm_table_supports_write_same (struct dm_table * t )
1529
1530
{
1530
1531
struct dm_target * ti ;
1531
- unsigned i = 0 ;
1532
+ unsigned i ;
1532
1533
1533
- while (i < dm_table_get_num_targets (t )) {
1534
- ti = dm_table_get_target (t , i ++ );
1534
+ for (i = 0 ; i < dm_table_get_num_targets (t ); i ++ ) {
1535
+ ti = dm_table_get_target (t , i );
1535
1536
1536
1537
if (!ti -> num_write_same_bios )
1537
1538
return false;
@@ -1555,7 +1556,7 @@ static int device_discard_capable(struct dm_target *ti, struct dm_dev *dev,
1555
1556
static bool dm_table_supports_discards (struct dm_table * t )
1556
1557
{
1557
1558
struct dm_target * ti ;
1558
- unsigned i = 0 ;
1559
+ unsigned i ;
1559
1560
1560
1561
/*
1561
1562
* Unless any target used by the table set discards_supported,
@@ -1564,8 +1565,8 @@ static bool dm_table_supports_discards(struct dm_table *t)
1564
1565
* so we need to use iterate_devices here, which targets
1565
1566
* supporting discard selectively must provide.
1566
1567
*/
1567
- while (i < dm_table_get_num_targets (t )) {
1568
- ti = dm_table_get_target (t , i ++ );
1568
+ for (i = 0 ; i < dm_table_get_num_targets (t ); i ++ ) {
1569
+ ti = dm_table_get_target (t , i );
1569
1570
1570
1571
if (!ti -> num_discard_bios )
1571
1572
continue ;
0 commit comments