Lines Matching +full:0 +full:x2002
40 for (i = 0; i < size; i++) in get_changed_bytes()
116 { .cache = REGCACHE_FLAT, .from_reg = 0 },
117 { .cache = REGCACHE_FLAT, .from_reg = 0, .fast_io = true },
118 { .cache = REGCACHE_FLAT, .from_reg = 0x2001 },
119 { .cache = REGCACHE_FLAT, .from_reg = 0x2002 },
120 { .cache = REGCACHE_FLAT, .from_reg = 0x2003 },
121 { .cache = REGCACHE_FLAT, .from_reg = 0x2004 },
122 { .cache = REGCACHE_RBTREE, .from_reg = 0 },
123 { .cache = REGCACHE_RBTREE, .from_reg = 0, .fast_io = true },
124 { .cache = REGCACHE_RBTREE, .from_reg = 0x2001 },
125 { .cache = REGCACHE_RBTREE, .from_reg = 0x2002 },
126 { .cache = REGCACHE_RBTREE, .from_reg = 0x2003 },
127 { .cache = REGCACHE_RBTREE, .from_reg = 0x2004 },
128 { .cache = REGCACHE_MAPLE, .from_reg = 0 },
129 { .cache = REGCACHE_MAPLE, .from_reg = 0, .fast_io = true },
130 { .cache = REGCACHE_MAPLE, .from_reg = 0x2001 },
131 { .cache = REGCACHE_MAPLE, .from_reg = 0x2002 },
132 { .cache = REGCACHE_MAPLE, .from_reg = 0x2003 },
133 { .cache = REGCACHE_MAPLE, .from_reg = 0x2004 },
139 { .cache = REGCACHE_RBTREE, .from_reg = 0 },
140 { .cache = REGCACHE_RBTREE, .from_reg = 0, .fast_io = true },
141 { .cache = REGCACHE_RBTREE, .from_reg = 0x2001 },
142 { .cache = REGCACHE_RBTREE, .from_reg = 0x2002 },
143 { .cache = REGCACHE_RBTREE, .from_reg = 0x2003 },
144 { .cache = REGCACHE_RBTREE, .from_reg = 0x2004 },
145 { .cache = REGCACHE_MAPLE, .from_reg = 0 },
146 { .cache = REGCACHE_MAPLE, .from_reg = 0, .fast_io = true },
147 { .cache = REGCACHE_MAPLE, .from_reg = 0x2001 },
148 { .cache = REGCACHE_MAPLE, .from_reg = 0x2002 },
149 { .cache = REGCACHE_MAPLE, .from_reg = 0x2003 },
150 { .cache = REGCACHE_MAPLE, .from_reg = 0x2004 },
170 if (config->max_register == 0) { in gen_regmap()
201 for (i = 0; i < config->num_reg_defaults; i++) { in gen_regmap()
250 KUNIT_EXPECT_EQ(test, 0, regmap_write(map, 0, val)); in basic_read_write()
251 KUNIT_EXPECT_EQ(test, 0, regmap_read(map, 0, &rval)); in basic_read_write()
255 KUNIT_EXPECT_EQ(test, config.cache_type == REGCACHE_NONE, data->read[0]); in basic_read_write()
279 KUNIT_EXPECT_EQ(test, 0, regmap_bulk_write(map, 0, val, in bulk_write()
281 for (i = 0; i < BLOCK_TEST_SIZE; i++) in bulk_write()
282 KUNIT_EXPECT_EQ(test, 0, regmap_read(map, i, &rval[i])); in bulk_write()
287 for (i = 0; i < BLOCK_TEST_SIZE; i++) in bulk_write()
309 for (i = 0; i < BLOCK_TEST_SIZE; i++) in bulk_read()
310 KUNIT_EXPECT_EQ(test, 0, regmap_write(map, i, val[i])); in bulk_read()
311 KUNIT_EXPECT_EQ(test, 0, regmap_bulk_read(map, 0, rval, in bulk_read()
316 for (i = 0; i < BLOCK_TEST_SIZE; i++) in bulk_read()
342 for (i = 0; i < BLOCK_TEST_SIZE; i++) { in multi_write()
345 sequence[i].delay_us = 0; in multi_write()
347 KUNIT_EXPECT_EQ(test, 0, in multi_write()
349 for (i = 0; i < BLOCK_TEST_SIZE; i++) in multi_write()
350 KUNIT_EXPECT_EQ(test, 0, regmap_read(map, i, &rval[i])); in multi_write()
355 for (i = 0; i < BLOCK_TEST_SIZE; i++) in multi_write()
378 for (i = 0; i < BLOCK_TEST_SIZE; i++) { in multi_read()
380 KUNIT_EXPECT_EQ(test, 0, regmap_write(map, i, val[i])); in multi_read()
382 KUNIT_EXPECT_EQ(test, 0, in multi_read()
387 for (i = 0; i < BLOCK_TEST_SIZE; i++) in multi_read()
412 KUNIT_EXPECT_EQ(test, 0, regmap_bulk_write(map, param->from_reg, val, ARRAY_SIZE(val))); in read_bypassed()
420 for (i = 0; i < ARRAY_SIZE(val); i++) { in read_bypassed()
423 KUNIT_EXPECT_EQ(test, 0, regmap_read(map, param->from_reg + i, &rval)); in read_bypassed()
427 KUNIT_EXPECT_EQ(test, 0, regmap_read_bypassed(map, param->from_reg + i, &rval)); in read_bypassed()
437 for (i = 0; i < ARRAY_SIZE(val); i++) { in read_bypassed()
442 for (i = 0; i < ARRAY_SIZE(val); i++) { in read_bypassed()
444 KUNIT_EXPECT_EQ(test, 0, regmap_read(map, param->from_reg + i, &rval)); in read_bypassed()
448 KUNIT_EXPECT_EQ(test, 0, regmap_read_bypassed(map, param->from_reg + i, &rval)); in read_bypassed()
478 KUNIT_EXPECT_EQ(test, 0, regmap_bulk_write(map, param->from_reg, val, ARRAY_SIZE(val))); in read_bypassed_volatile()
486 for (i = 0; i < ARRAY_SIZE(val); i++) { in read_bypassed_volatile()
488 KUNIT_EXPECT_EQ(test, (i == 5) ? 0 : -EBUSY, in read_bypassed_volatile()
493 KUNIT_EXPECT_EQ(test, 0, regmap_read_bypassed(map, param->from_reg + i, &rval)); in read_bypassed_volatile()
503 for (i = 0; i < ARRAY_SIZE(val); i++) { in read_bypassed_volatile()
508 for (i = 0; i < ARRAY_SIZE(val); i++) { in read_bypassed_volatile()
513 KUNIT_EXPECT_EQ(test, 0, regmap_read_bypassed(map, param->from_reg + i, &rval)); in read_bypassed_volatile()
539 for (i = 0; i < BLOCK_TEST_SIZE; i++) in write_readonly()
543 for (i = 0; i < BLOCK_TEST_SIZE; i++) in write_readonly()
544 KUNIT_EXPECT_EQ(test, i != 5, regmap_write(map, i, val) == 0); in write_readonly()
547 for (i = 0; i < BLOCK_TEST_SIZE; i++) in write_readonly()
567 for (i = 0; i < BLOCK_TEST_SIZE; i++) in read_writeonly()
574 for (i = 0; i < BLOCK_TEST_SIZE; i++) { in read_writeonly()
577 regmap_read(map, i, &val) == 0); in read_writeonly()
579 KUNIT_EXPECT_EQ(test, 0, regmap_read(map, i, &val)); in read_writeonly()
604 KUNIT_EXPECT_EQ(test, 0, regmap_bulk_read(map, 0, rval, in reg_defaults()
609 for (i = 0; i < BLOCK_TEST_SIZE; i++) in reg_defaults()
630 for (i = 0; i < BLOCK_TEST_SIZE; i++) { in reg_defaults_read_dev()
636 KUNIT_EXPECT_EQ(test, 0, regmap_bulk_read(map, 0, rval, in reg_defaults_read_dev()
641 for (i = 0; i < BLOCK_TEST_SIZE; i++) in reg_defaults_read_dev()
664 KUNIT_EXPECT_EQ(test, 0, regmap_bulk_read(map, 0, rval, in register_patch()
668 patch[0].reg = 2; in register_patch()
669 patch[0].def = rval[2] + 1; in register_patch()
670 patch[0].delay_us = 0; in register_patch()
673 patch[1].delay_us = 0; in register_patch()
674 KUNIT_EXPECT_EQ(test, 0, regmap_register_patch(map, patch, in register_patch()
678 for (i = 0; i < BLOCK_TEST_SIZE; i++) { in register_patch()
718 for (i = 0; i < BLOCK_TEST_SIZE; i++) { in stride()
723 KUNIT_EXPECT_NE(test, 0, regmap_read(map, i, &rval)); in stride()
724 KUNIT_EXPECT_NE(test, 0, regmap_write(map, i, rval)); in stride()
728 KUNIT_EXPECT_EQ(test, 0, regmap_read(map, i, &rval)); in stride()
733 KUNIT_EXPECT_EQ(test, 0, regmap_write(map, i, rval)); in stride()
741 .selector_mask = 0xff,
795 KUNIT_EXPECT_EQ(test, 0, regmap_write(map, test_range.selector_reg, in basic_ranges()
801 KUNIT_EXPECT_EQ(test, 0, regmap_write(map, test_range.range_min, 0)); in basic_ranges()
807 KUNIT_EXPECT_EQ(test, 0, regmap_write(map, in basic_ranges()
810 0)); in basic_ranges()
817 KUNIT_EXPECT_EQ(test, 0, regmap_read(map, test_range.range_min, &val)); in basic_ranges()
823 KUNIT_EXPECT_EQ(test, 0, regmap_read(map, in basic_ranges()
862 for (i = 0; i < config.max_register; i += 100) in stress_insert()
863 KUNIT_EXPECT_EQ(test, 0, regmap_write(map, i, vals[i])); in stress_insert()
864 for (i = 0; i < config.max_register; i += 50) in stress_insert()
865 KUNIT_EXPECT_EQ(test, 0, regmap_write(map, i, vals[i])); in stress_insert()
866 for (i = 0; i < config.max_register; i += 25) in stress_insert()
867 KUNIT_EXPECT_EQ(test, 0, regmap_write(map, i, vals[i])); in stress_insert()
868 for (i = 0; i < config.max_register; i += 10) in stress_insert()
869 KUNIT_EXPECT_EQ(test, 0, regmap_write(map, i, vals[i])); in stress_insert()
870 for (i = 0; i < config.max_register; i += 5) in stress_insert()
871 KUNIT_EXPECT_EQ(test, 0, regmap_write(map, i, vals[i])); in stress_insert()
872 for (i = 0; i < config.max_register; i += 3) in stress_insert()
873 KUNIT_EXPECT_EQ(test, 0, regmap_write(map, i, vals[i])); in stress_insert()
874 for (i = 0; i < config.max_register; i += 2) in stress_insert()
875 KUNIT_EXPECT_EQ(test, 0, regmap_write(map, i, vals[i])); in stress_insert()
876 for (i = 0; i < config.max_register; i++) in stress_insert()
877 KUNIT_EXPECT_EQ(test, 0, regmap_write(map, i, vals[i])); in stress_insert()
880 for (i = 0; i < config.max_register; i ++) { in stress_insert()
881 KUNIT_EXPECT_EQ(test, 0, regmap_read(map, i, &rval)); in stress_insert()
905 KUNIT_EXPECT_EQ(test, 0, regmap_write(map, param->from_reg, val)); in cache_bypass()
909 KUNIT_EXPECT_EQ(test, 0, regmap_write(map, param->from_reg, val + 1)); in cache_bypass()
912 KUNIT_EXPECT_EQ(test, 0, regmap_read(map, param->from_reg, &rval)); in cache_bypass()
918 KUNIT_EXPECT_EQ(test, 0, regmap_read(map, param->from_reg, &rval)); in cache_bypass()
941 KUNIT_EXPECT_EQ(test, 0, regmap_bulk_write(map, param->from_reg, val, in cache_sync_marked_dirty()
943 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_sync_marked_dirty()
948 memset(data->vals, 0, sizeof(val)); in cache_sync_marked_dirty()
949 KUNIT_EXPECT_EQ(test, 0, regcache_sync(map)); in cache_sync_marked_dirty()
953 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_sync_marked_dirty()
974 val_mask = GENMASK(config.val_bits - 1, 0); in cache_sync_after_cache_only()
978 KUNIT_EXPECT_EQ(test, 0, regmap_bulk_write(map, param->from_reg, val, in cache_sync_after_cache_only()
980 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_sync_after_cache_only()
985 for (i = 0; i < ARRAY_SIZE(val); ++i) in cache_sync_after_cache_only()
988 KUNIT_EXPECT_EQ(test, 0, regmap_bulk_write(map, param->from_reg, val, in cache_sync_after_cache_only()
990 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_sync_after_cache_only()
998 KUNIT_EXPECT_EQ(test, 0, regcache_sync(map)); in cache_sync_after_cache_only()
1002 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_sync_after_cache_only()
1026 KUNIT_EXPECT_EQ(test, 0, regmap_write(map, param->from_reg + 2, val)); in cache_sync_defaults_marked_dirty()
1030 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_sync_defaults_marked_dirty()
1032 KUNIT_EXPECT_EQ(test, 0, regcache_sync(map)); in cache_sync_defaults_marked_dirty()
1035 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_sync_defaults_marked_dirty()
1039 for (i = 0; i < config.num_reg_defaults; ++i) in cache_sync_defaults_marked_dirty()
1040 KUNIT_EXPECT_EQ(test, 0, regmap_write(map, config.reg_defaults[i].reg, in cache_sync_defaults_marked_dirty()
1047 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_sync_defaults_marked_dirty()
1050 KUNIT_EXPECT_EQ(test, 0, regcache_sync(map)); in cache_sync_defaults_marked_dirty()
1051 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_sync_defaults_marked_dirty()
1072 KUNIT_EXPECT_EQ(test, 0, regmap_read(map, param->from_reg + 2, &orig_val)); in cache_sync_default_after_cache_only()
1076 KUNIT_EXPECT_EQ(test, 0, regmap_write(map, param->from_reg + 2, orig_val + 1)); in cache_sync_default_after_cache_only()
1080 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_sync_default_after_cache_only()
1082 KUNIT_EXPECT_EQ(test, 0, regcache_sync(map)); in cache_sync_default_after_cache_only()
1090 KUNIT_EXPECT_EQ(test, 0, regmap_write(map, param->from_reg + 2, orig_val)); in cache_sync_default_after_cache_only()
1094 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_sync_default_after_cache_only()
1097 KUNIT_EXPECT_EQ(test, 0, regcache_sync(map)); in cache_sync_default_after_cache_only()
1120 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_sync_readonly()
1121 KUNIT_EXPECT_EQ(test, 0, regmap_read(map, param->from_reg + i, &val)); in cache_sync_readonly()
1126 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_sync_readonly()
1127 KUNIT_EXPECT_EQ(test, i != 5, regmap_write(map, param->from_reg + i, val) == 0); in cache_sync_readonly()
1131 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_sync_readonly()
1133 KUNIT_EXPECT_EQ(test, 0, regcache_sync(map)); in cache_sync_readonly()
1136 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_sync_readonly()
1160 KUNIT_EXPECT_EQ(test, 0, regmap_bulk_read(map, param->from_reg, rval, in cache_sync_patch()
1164 patch[0].reg = param->from_reg + 2; in cache_sync_patch()
1165 patch[0].def = rval[2] + 1; in cache_sync_patch()
1166 patch[0].delay_us = 0; in cache_sync_patch()
1169 patch[1].delay_us = 0; in cache_sync_patch()
1170 KUNIT_EXPECT_EQ(test, 0, regmap_register_patch(map, patch, in cache_sync_patch()
1175 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_sync_patch()
1177 KUNIT_EXPECT_EQ(test, 0, regcache_sync(map)); in cache_sync_patch()
1180 for (i = 0; i < BLOCK_TEST_SIZE; i++) { in cache_sync_patch()
1181 KUNIT_EXPECT_EQ(test, 0, regmap_read(map, param->from_reg + i, &val)); in cache_sync_patch()
1216 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_drop()
1218 KUNIT_EXPECT_EQ(test, 0, regmap_bulk_read(map, param->from_reg, rval, in cache_drop()
1220 for (i = 0; i < BLOCK_TEST_SIZE; i++) { in cache_drop()
1227 KUNIT_EXPECT_EQ(test, 0, regcache_drop_region(map, param->from_reg + 3, in cache_drop()
1231 KUNIT_EXPECT_EQ(test, 0, regmap_bulk_read(map, param->from_reg, rval, in cache_drop()
1233 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_drop()
1259 for (i = 0; i < config.max_register + 1; i++) in cache_drop_with_non_contiguous_ranges()
1264 for (rangeidx = 0; rangeidx < num_ranges; rangeidx += 2) { in cache_drop_with_non_contiguous_ranges()
1266 KUNIT_EXPECT_EQ(test, 0, regmap_bulk_write(map, reg, in cache_drop_with_non_contiguous_ranges()
1276 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_drop_with_non_contiguous_ranges()
1282 KUNIT_EXPECT_EQ(test, 0, regcache_drop_region(map, reg, reg + BLOCK_TEST_SIZE - 1)); in cache_drop_with_non_contiguous_ranges()
1286 KUNIT_EXPECT_EQ(test, 0, regcache_drop_region(map, reg + 3, reg + 5)); in cache_drop_with_non_contiguous_ranges()
1288 /* Mark dirty and reset mock registers to 0 */ in cache_drop_with_non_contiguous_ranges()
1290 for (i = 0; i < config.max_register + 1; i++) { in cache_drop_with_non_contiguous_ranges()
1291 data->vals[i] = 0; in cache_drop_with_non_contiguous_ranges()
1295 /* The registers that were dropped from range 4 should now remain at 0 */ in cache_drop_with_non_contiguous_ranges()
1296 val[4 / 2][3] = 0; in cache_drop_with_non_contiguous_ranges()
1297 val[4 / 2][4] = 0; in cache_drop_with_non_contiguous_ranges()
1298 val[4 / 2][5] = 0; in cache_drop_with_non_contiguous_ranges()
1301 KUNIT_EXPECT_EQ(test, 0, regcache_sync(map)); in cache_drop_with_non_contiguous_ranges()
1306 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_drop_with_non_contiguous_ranges()
1311 for (rangeidx = 0; rangeidx < num_ranges; rangeidx += 2) { in cache_drop_with_non_contiguous_ranges()
1316 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_drop_with_non_contiguous_ranges()
1325 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_drop_with_non_contiguous_ranges()
1330 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_drop_with_non_contiguous_ranges()
1336 for (i = 0; i < param->from_reg; i++) in cache_drop_with_non_contiguous_ranges()
1358 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_drop_all_and_sync_marked_dirty()
1360 KUNIT_EXPECT_EQ(test, 0, regmap_bulk_read(map, param->from_reg, rval, in cache_drop_all_and_sync_marked_dirty()
1365 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_drop_all_and_sync_marked_dirty()
1366 KUNIT_EXPECT_EQ(test, 0, regmap_write(map, param->from_reg + i, rval[i] + 1)); in cache_drop_all_and_sync_marked_dirty()
1369 KUNIT_EXPECT_EQ(test, 0, regcache_drop_region(map, 0, config.max_register)); in cache_drop_all_and_sync_marked_dirty()
1373 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_drop_all_and_sync_marked_dirty()
1376 KUNIT_EXPECT_EQ(test, 0, regcache_sync(map)); in cache_drop_all_and_sync_marked_dirty()
1377 for (i = 0; i <= config.max_register; i++) in cache_drop_all_and_sync_marked_dirty()
1398 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_drop_all_and_sync_no_defaults()
1400 KUNIT_EXPECT_EQ(test, 0, regmap_bulk_read(map, param->from_reg, rval, in cache_drop_all_and_sync_no_defaults()
1405 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_drop_all_and_sync_no_defaults()
1406 KUNIT_EXPECT_EQ(test, 0, regmap_write(map, param->from_reg + i, rval[i] + 1)); in cache_drop_all_and_sync_no_defaults()
1409 KUNIT_EXPECT_EQ(test, 0, regcache_drop_region(map, 0, config.max_register)); in cache_drop_all_and_sync_no_defaults()
1415 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_drop_all_and_sync_no_defaults()
1418 KUNIT_EXPECT_EQ(test, 0, regcache_sync(map)); in cache_drop_all_and_sync_no_defaults()
1419 for (i = 0; i <= config.max_register; i++) in cache_drop_all_and_sync_no_defaults()
1441 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_drop_all_and_sync_has_defaults()
1443 KUNIT_EXPECT_EQ(test, 0, regmap_bulk_read(map, param->from_reg, rval, in cache_drop_all_and_sync_has_defaults()
1448 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_drop_all_and_sync_has_defaults()
1449 KUNIT_EXPECT_EQ(test, 0, regmap_write(map, param->from_reg + i, rval[i] + 1)); in cache_drop_all_and_sync_has_defaults()
1452 KUNIT_EXPECT_EQ(test, 0, regcache_drop_region(map, 0, config.max_register)); in cache_drop_all_and_sync_has_defaults()
1458 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_drop_all_and_sync_has_defaults()
1461 KUNIT_EXPECT_EQ(test, 0, regcache_sync(map)); in cache_drop_all_and_sync_has_defaults()
1462 for (i = 0; i <= config.max_register; i++) in cache_drop_all_and_sync_has_defaults()
1482 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_present()
1486 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_present()
1490 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_present()
1494 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_present()
1495 KUNIT_EXPECT_EQ(test, 0, regmap_read(map, param->from_reg + i, &val)); in cache_present()
1498 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_present()
1518 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_write_zero()
1522 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_write_zero()
1526 for (i = 0; i < BLOCK_TEST_SIZE; i++) in cache_write_zero()
1530 KUNIT_EXPECT_EQ(test, 0, regmap_write(map, 1, 0)); in cache_write_zero()
1533 KUNIT_EXPECT_EQ(test, 0, regmap_read(map, 1, &val)); in cache_write_zero()
1534 KUNIT_EXPECT_EQ(test, 0, val); in cache_write_zero()
1540 KUNIT_EXPECT_EQ(test, 0, regcache_drop_region(map, 1, 1)); in cache_write_zero()
1566 KUNIT_ASSERT_EQ(test, 0, regmap_write(map, i, 0)); in cache_range_window_reg()
1572 KUNIT_ASSERT_EQ(test, 0, regmap_write(map, test_range.range_min, 0)); in cache_range_window_reg()
1574 KUNIT_ASSERT_EQ(test, val, 0); in cache_range_window_reg()
1578 KUNIT_ASSERT_EQ(test, 0, regcache_sync(map)); in cache_range_window_reg()
1581 KUNIT_ASSERT_EQ(test, 0, regmap_write(map, test_range.range_min, 0)); in cache_range_window_reg()
1583 KUNIT_ASSERT_EQ(test, val, 0); in cache_range_window_reg()
1587 KUNIT_ASSERT_EQ(test, 0, regcache_sync(map)); in cache_range_window_reg()
1590 KUNIT_ASSERT_EQ(test, 0, regmap_write(map, test_range.range_max, 0)); in cache_range_window_reg()
1665 for (i = 0; i < config->num_reg_defaults; i++) { in gen_raw_regmap()
1685 config->num_reg_defaults = 0; in gen_raw_regmap()
1721 for (i = 0; i < config.max_register + 1; i++) { in raw_read_defaults_single()
1722 KUNIT_EXPECT_EQ(test, 0, regmap_read(map, i, &rval)); in raw_read_defaults_single()
1751 KUNIT_EXPECT_EQ(test, 0, regmap_raw_read(map, 0, rval, val_len)); in raw_read_defaults()
1752 for (i = 0; i < config.max_register + 1; i++) { in raw_read_defaults()
1780 KUNIT_EXPECT_EQ(test, 0, regmap_write(map, 0, val)); in raw_write_read_single()
1781 KUNIT_EXPECT_EQ(test, 0, regmap_read(map, 0, &rval)); in raw_write_read_single()
1807 KUNIT_EXPECT_EQ(test, 0, regmap_raw_write(map, 2, val, sizeof(val))); in raw_write()
1810 for (i = 0; i < config.max_register + 1; i++) { in raw_write()
1811 KUNIT_EXPECT_EQ(test, 0, regmap_read(map, i, &rval)); in raw_write()
1836 return reg == 0; in reg_zero()
1841 return reg == 0; in ram_reg_zero()
1876 KUNIT_EXPECT_EQ(test, 0, regmap_write(map, 1, val_test)); in raw_noinc_write()
1879 KUNIT_EXPECT_EQ(test, 0, regmap_noinc_write(map, 0, val_array, in raw_noinc_write()
1883 KUNIT_EXPECT_EQ(test, 0, regmap_read(map, 0, &val)); in raw_noinc_write()
1887 KUNIT_EXPECT_EQ(test, 0, regmap_read(map, 1, &val)); in raw_noinc_write()
1910 get_changed_bytes(&hw_buf[2], &val[0], sizeof(val)); in raw_sync()
1914 KUNIT_EXPECT_EQ(test, 0, regmap_raw_write(map, 2, val, in raw_sync()
1916 KUNIT_EXPECT_EQ(test, 0, regmap_write(map, 4, val[2])); in raw_sync()
1919 for (i = 0; i < config.max_register + 1; i++) { in raw_sync()
1920 KUNIT_EXPECT_EQ(test, 0, regmap_read(map, i, &rval)); in raw_sync()
1952 KUNIT_EXPECT_MEMNEQ(test, &hw_buf[2], &val[0], sizeof(val)); in raw_sync()
1954 for (i = 0; i < config.max_register + 1; i++) in raw_sync()
1960 KUNIT_EXPECT_EQ(test, 0, regcache_sync(map)); in raw_sync()
1963 KUNIT_EXPECT_MEMEQ(test, &hw_buf[2], &val[0], sizeof(val)); in raw_sync()
1986 KUNIT_EXPECT_EQ(test, 0, regmap_write(map, test_range.selector_reg, in raw_ranges()
1992 KUNIT_EXPECT_EQ(test, 0, regmap_write(map, test_range.range_min, 0)); in raw_ranges()
1998 KUNIT_EXPECT_EQ(test, 0, regmap_write(map, in raw_ranges()
2001 0)); in raw_ranges()
2008 KUNIT_EXPECT_EQ(test, 0, regmap_read(map, test_range.range_min, &val)); in raw_ranges()
2014 KUNIT_EXPECT_EQ(test, 0, regmap_read(map, in raw_ranges()
2088 return 0; in regmap_test_init()