Hi Andrey, > > > > When KASAN is configured in store-only mode, > > fetch/load operations do not trigger tag check faults. > > As a result, the outcome of some test cases may differ > > compared to when KASAN is configured without store-only mode. > > > > To address this: > > 1. Replace fetch/load expressions that would > > normally trigger tag check faults with store operation > > when running under store-only and sync mode. > > In case of async/asymm mode, skip the store operation triggering > > tag check fault since it corrupts memory. > > > > 2. Skip some testcases affected by initial value > > (i.e) atomic_cmpxchg() testcase maybe successd if > > it passes valid atomic_t address and invalid oldaval address. > > In this case, if invalid atomic_t doesn't have the same oldval, > > it won't trigger store operation so the test will pass. > > > > Signed-off-by: Yeoreum Yun <yeoreum.yun@xxxxxxx> > > --- > > mm/kasan/kasan_test_c.c | 423 ++++++++++++++++++++++++++++++++-------- > > 1 file changed, 341 insertions(+), 82 deletions(-) > > > > diff --git a/mm/kasan/kasan_test_c.c b/mm/kasan/kasan_test_c.c > > index 2aa12dfa427a..22d5d6d6cd9f 100644 > > --- a/mm/kasan/kasan_test_c.c > > +++ b/mm/kasan/kasan_test_c.c > > @@ -94,11 +94,13 @@ static void kasan_test_exit(struct kunit *test) > > } > > > > /** > > - * KUNIT_EXPECT_KASAN_FAIL - check that the executed expression produces a > > - * KASAN report; causes a KUnit test failure otherwise. > > + * _KUNIT_EXPECT_KASAN_TEMPLATE - check that the executed expression produces > > + * a KASAN report or not; a KUnit test failure when it's different from @produce. > > * > > * @test: Currently executing KUnit test. > > - * @expression: Expression that must produce a KASAN report. > > + * @expr: Expression produce a KASAN report or not. > > + * @expr_str: Expression string > > + * @produce: expression should produce a KASAN report. > > * > > * For hardware tag-based KASAN, when a synchronous tag fault happens, tag > > * checking is auto-disabled. When this happens, this test handler reenables > > @@ -110,25 +112,29 @@ static void kasan_test_exit(struct kunit *test) > > * Use READ/WRITE_ONCE() for the accesses and compiler barriers around the > > * expression to prevent that. > > * > > - * In between KUNIT_EXPECT_KASAN_FAIL checks, test_status.report_found is kept > > + * In between _KUNIT_EXPECT_KASAN_TEMPLATE checks, test_status.report_found is kept > > * as false. This allows detecting KASAN reports that happen outside of the > > * checks by asserting !test_status.report_found at the start of > > - * KUNIT_EXPECT_KASAN_FAIL and in kasan_test_exit. > > + * _KUNIT_EXPECT_KASAN_TEMPLATE and in kasan_test_exit. > > */ > > -#define KUNIT_EXPECT_KASAN_FAIL(test, expression) do { \ > > +#define _KUNIT_EXPECT_KASAN_TEMPLATE(test, expr, expr_str, produce) \ > > +do { \ > > if (IS_ENABLED(CONFIG_KASAN_HW_TAGS) && \ > > kasan_sync_fault_possible()) \ > > migrate_disable(); \ > > KUNIT_EXPECT_FALSE(test, READ_ONCE(test_status.report_found)); \ > > barrier(); \ > > - expression; \ > > + expr; \ > > barrier(); \ > > if (kasan_async_fault_possible()) \ > > kasan_force_async_fault(); \ > > - if (!READ_ONCE(test_status.report_found)) { \ > > - KUNIT_FAIL(test, KUNIT_SUBTEST_INDENT "KASAN failure " \ > > - "expected in \"" #expression \ > > - "\", but none occurred"); \ > > + if (READ_ONCE(test_status.report_found) != produce) { \ > > + KUNIT_FAIL(test, KUNIT_SUBTEST_INDENT "KASAN %s " \ > > + "expected in \"" expr_str \ > > + "\", but %soccurred", \ > > + (produce ? "failure" : "success"), \ > > + (test_status.report_found ? \ > > + "" : "none ")); \ > > } \ > > if (IS_ENABLED(CONFIG_KASAN_HW_TAGS) && \ > > kasan_sync_fault_possible()) { \ > > @@ -141,6 +147,26 @@ static void kasan_test_exit(struct kunit *test) > > WRITE_ONCE(test_status.async_fault, false); \ > > } while (0) > > > > +/* > > + * KUNIT_EXPECT_KASAN_FAIL - check that the executed expression produces a > > + * KASAN report; causes a KUnit test failure otherwise. > > + * > > + * @test: Currently executing KUnit test. > > + * @expr: Expression produce a KASAN report. > > + */ > > +#define KUNIT_EXPECT_KASAN_FAIL(test, expr) \ > > + _KUNIT_EXPECT_KASAN_TEMPLATE(test, expr, #expr, true) > > + > > +/* > > + * KUNIT_EXPECT_KASAN_SUCCESS - check that the executed expression doesn't > > + * produces a KASAN report; causes a KUnit test failure otherwise. > > + * > > + * @test: Currently executing KUnit test. > > + * @expr: Expression doesn't produce a KASAN report. > > + */ > > +#define KUNIT_EXPECT_KASAN_SUCCESS(test, expr) \ > > + _KUNIT_EXPECT_KASAN_TEMPLATE(test, expr, #expr, false) > > + > > #define KASAN_TEST_NEEDS_CONFIG_ON(test, config) do { \ > > if (!IS_ENABLED(config)) \ > > kunit_skip((test), "Test requires " #config "=y"); \ > > @@ -183,8 +209,15 @@ static void kmalloc_oob_right(struct kunit *test) > > KUNIT_EXPECT_KASAN_FAIL(test, ptr[size + 5] = 'y'); > > > > /* Out-of-bounds access past the aligned kmalloc object. */ > > - KUNIT_EXPECT_KASAN_FAIL(test, ptr[0] = > > - ptr[size + KASAN_GRANULE_SIZE + 5]); > > + if (kasan_stonly_enabled()) { > > + KUNIT_EXPECT_KASAN_SUCCESS(test, ptr[0] = > > + ptr[size + KASAN_GRANULE_SIZE + 5]); > > + if (!kasan_async_fault_possible()) > > + KUNIT_EXPECT_KASAN_FAIL(test, > > + ptr[size + KASAN_GRANULE_SIZE + 5] = ptr[0]); > > + } else > > + KUNIT_EXPECT_KASAN_FAIL(test, ptr[0] = > > + ptr[size + KASAN_GRANULE_SIZE + 5]); > > > > kfree(ptr); > > } > > @@ -198,7 +231,13 @@ static void kmalloc_oob_left(struct kunit *test) > > KUNIT_ASSERT_NOT_ERR_OR_NULL(test, ptr); > > > > OPTIMIZER_HIDE_VAR(ptr); > > - KUNIT_EXPECT_KASAN_FAIL(test, *ptr = *(ptr - 1)); > > + if (kasan_stonly_enabled()) { > > + KUNIT_EXPECT_KASAN_SUCCESS(test, *ptr = *(ptr - 1)); > > + if (!kasan_async_fault_possible()) > > + KUNIT_EXPECT_KASAN_FAIL(test, *(ptr - 1) = *(ptr)); > > + } else > > + KUNIT_EXPECT_KASAN_FAIL(test, *ptr = *(ptr - 1)); > > + > > kfree(ptr); > > } > > > > @@ -211,7 +250,13 @@ static void kmalloc_node_oob_right(struct kunit *test) > > KUNIT_ASSERT_NOT_ERR_OR_NULL(test, ptr); > > > > OPTIMIZER_HIDE_VAR(ptr); > > - KUNIT_EXPECT_KASAN_FAIL(test, ptr[0] = ptr[size]); > > + if (kasan_stonly_enabled()) { > > + KUNIT_EXPECT_KASAN_SUCCESS(test, ptr[0] = ptr[size]); > > + if (!kasan_async_fault_possible()) > > + KUNIT_EXPECT_KASAN_FAIL(test, ptr[size] = ptr[0]); > > + } else > > + KUNIT_EXPECT_KASAN_FAIL(test, ptr[0] = ptr[size]); > > + > > kfree(ptr); > > } > > > > @@ -291,7 +336,12 @@ static void kmalloc_large_uaf(struct kunit *test) > > KUNIT_ASSERT_NOT_ERR_OR_NULL(test, ptr); > > kfree(ptr); > > > > - KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[0]); > > + if (kasan_stonly_enabled()) { > > + KUNIT_EXPECT_KASAN_SUCCESS(test, ((volatile char *)ptr)[0]); > > + if (!kasan_async_fault_possible()) > > + KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[0] = 0); > > + } else > > + KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[0]); > > } > > > > static void kmalloc_large_invalid_free(struct kunit *test) > > @@ -323,7 +373,13 @@ static void page_alloc_oob_right(struct kunit *test) > > ptr = page_address(pages); > > KUNIT_ASSERT_NOT_ERR_OR_NULL(test, ptr); > > > > - KUNIT_EXPECT_KASAN_FAIL(test, ptr[0] = ptr[size]); > > + if (kasan_stonly_enabled()) { > > + KUNIT_EXPECT_KASAN_SUCCESS(test, ptr[0] = ptr[size]); > > + if (!kasan_async_fault_possible()) > > + KUNIT_EXPECT_KASAN_FAIL(test, ptr[size] = ptr[0]); > > + } else > > + KUNIT_EXPECT_KASAN_FAIL(test, ptr[0] = ptr[size]); > > + > > free_pages((unsigned long)ptr, order); > > } > > > > @@ -338,7 +394,12 @@ static void page_alloc_uaf(struct kunit *test) > > KUNIT_ASSERT_NOT_ERR_OR_NULL(test, ptr); > > free_pages((unsigned long)ptr, order); > > > > - KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[0]); > > + if (kasan_stonly_enabled()) { > > + KUNIT_EXPECT_KASAN_SUCCESS(test, ((volatile char *)ptr)[0]); > > + if (!kasan_async_fault_possible()) > > + KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[0] = 0); > > + } else > > + KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[0]); > > } > > > > static void krealloc_more_oob_helper(struct kunit *test, > > @@ -455,10 +516,15 @@ static void krealloc_uaf(struct kunit *test) > > ptr1 = kmalloc(size1, GFP_KERNEL); > > KUNIT_ASSERT_NOT_ERR_OR_NULL(test, ptr1); > > kfree(ptr1); > > - > > KUNIT_EXPECT_KASAN_FAIL(test, ptr2 = krealloc(ptr1, size2, GFP_KERNEL)); > > KUNIT_ASSERT_NULL(test, ptr2); > > - KUNIT_EXPECT_KASAN_FAIL(test, *(volatile char *)ptr1); > > + > > + if (kasan_stonly_enabled()) { > > + KUNIT_EXPECT_KASAN_SUCCESS(test, *(volatile char *)ptr1); > > + if (!kasan_async_fault_possible()) > > + KUNIT_EXPECT_KASAN_FAIL(test, *(volatile char *)ptr1 = 0); > > + } else > > + KUNIT_EXPECT_KASAN_FAIL(test, *(volatile char *)ptr1); > > } > > > > static void kmalloc_oob_16(struct kunit *test) > > @@ -501,7 +567,13 @@ static void kmalloc_uaf_16(struct kunit *test) > > KUNIT_ASSERT_NOT_ERR_OR_NULL(test, ptr2); > > kfree(ptr2); > > > > - KUNIT_EXPECT_KASAN_FAIL(test, *ptr1 = *ptr2); > > + if (kasan_stonly_enabled()) { > > + KUNIT_EXPECT_KASAN_SUCCESS(test, *ptr1 = *ptr2); > > + if (!kasan_async_fault_possible()) > > + KUNIT_EXPECT_KASAN_FAIL(test, *ptr2 = *ptr1); > > + } else > > + KUNIT_EXPECT_KASAN_FAIL(test, *ptr1 = *ptr2); > > + > > kfree(ptr1); > > } > > > > @@ -640,8 +712,17 @@ static void kmalloc_memmove_invalid_size(struct kunit *test) > > memset((char *)ptr, 0, 64); > > OPTIMIZER_HIDE_VAR(ptr); > > OPTIMIZER_HIDE_VAR(invalid_size); > > - KUNIT_EXPECT_KASAN_FAIL(test, > > - memmove((char *)ptr, (char *)ptr + 4, invalid_size)); > > + > > + if (kasan_stonly_enabled()) { > > + KUNIT_EXPECT_KASAN_SUCCESS(test, > > + memmove((char *)ptr, (char *)ptr + 4, invalid_size)); > > + if (!kasan_async_fault_possible()) > > + KUNIT_EXPECT_KASAN_FAIL(test, > > + memmove((char *)ptr + 4, (char *)ptr, invalid_size)); > > + } else > > + KUNIT_EXPECT_KASAN_FAIL(test, > > + memmove((char *)ptr, (char *)ptr + 4, invalid_size)); > > + > > kfree(ptr); > > } > > > > @@ -654,7 +735,13 @@ static void kmalloc_uaf(struct kunit *test) > > KUNIT_ASSERT_NOT_ERR_OR_NULL(test, ptr); > > > > kfree(ptr); > > - KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[8]); > > + > > + if (kasan_stonly_enabled()) { > > + KUNIT_EXPECT_KASAN_SUCCESS(test, ((volatile char *)ptr)[8]); > > + if (!kasan_sync_fault_possible()) > > + KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[8] = 0); > > + } else > > + KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[8]); > > } > > > > static void kmalloc_uaf_memset(struct kunit *test) > > @@ -701,7 +788,13 @@ static void kmalloc_uaf2(struct kunit *test) > > goto again; > > } > > > > - KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr1)[40]); > > + if (kasan_stonly_enabled()) { > > + KUNIT_EXPECT_KASAN_SUCCESS(test, ((volatile char *)ptr1)[40]); > > + if (!kasan_sync_fault_possible()) > > + KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr1)[40] = 0); > > + } else > > + KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr1)[40]); > > + > > KUNIT_EXPECT_PTR_NE(test, ptr1, ptr2); > > > > kfree(ptr2); > > @@ -727,19 +820,35 @@ static void kmalloc_uaf3(struct kunit *test) > > KUNIT_ASSERT_NOT_ERR_OR_NULL(test, ptr2); > > kfree(ptr2); > > > > - KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr1)[8]); > > + if (kasan_stonly_enabled()) { > > + KUNIT_EXPECT_KASAN_SUCCESS(test, ((volatile char *)ptr1)[8]); > > + if (!kasan_sync_fault_possible()) > > + KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr1)[8] = 0); > > + } else > > + KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr1)[8]); > > } > > > > static void kasan_atomics_helper(struct kunit *test, void *unsafe, void *safe) > > { > > int *i_unsafe = unsafe; > > > > - KUNIT_EXPECT_KASAN_FAIL(test, READ_ONCE(*i_unsafe)); > > + if (kasan_stonly_enabled()) > > + KUNIT_EXPECT_KASAN_SUCCESS(test, READ_ONCE(*i_unsafe)); > > + else > > + KUNIT_EXPECT_KASAN_FAIL(test, READ_ONCE(*i_unsafe)); > > + > > KUNIT_EXPECT_KASAN_FAIL(test, WRITE_ONCE(*i_unsafe, 42)); > > - KUNIT_EXPECT_KASAN_FAIL(test, smp_load_acquire(i_unsafe)); > > + if (kasan_stonly_enabled()) > > + KUNIT_EXPECT_KASAN_SUCCESS(test, smp_load_acquire(i_unsafe)); > > + else > > + KUNIT_EXPECT_KASAN_FAIL(test, smp_load_acquire(i_unsafe)); > > KUNIT_EXPECT_KASAN_FAIL(test, smp_store_release(i_unsafe, 42)); > > > > - KUNIT_EXPECT_KASAN_FAIL(test, atomic_read(unsafe)); > > + if (kasan_stonly_enabled()) > > + KUNIT_EXPECT_KASAN_SUCCESS(test, atomic_read(unsafe)); > > + else > > + KUNIT_EXPECT_KASAN_FAIL(test, atomic_read(unsafe)); > > + > > KUNIT_EXPECT_KASAN_FAIL(test, atomic_set(unsafe, 42)); > > KUNIT_EXPECT_KASAN_FAIL(test, atomic_add(42, unsafe)); > > KUNIT_EXPECT_KASAN_FAIL(test, atomic_sub(42, unsafe)); > > @@ -752,18 +861,38 @@ static void kasan_atomics_helper(struct kunit *test, void *unsafe, void *safe) > > KUNIT_EXPECT_KASAN_FAIL(test, atomic_xchg(unsafe, 42)); > > KUNIT_EXPECT_KASAN_FAIL(test, atomic_cmpxchg(unsafe, 21, 42)); > > KUNIT_EXPECT_KASAN_FAIL(test, atomic_try_cmpxchg(unsafe, safe, 42)); > > - KUNIT_EXPECT_KASAN_FAIL(test, atomic_try_cmpxchg(safe, unsafe, 42)); > > + > > + /* > > + * The result of the test below may vary due to garbage values of unsafe in > > + * store-only mode. Therefore, skip this test when KASAN is configured > > + * in store-only mode. > > + */ > > + if (!kasan_stonly_enabled()) > > + KUNIT_EXPECT_KASAN_FAIL(test, atomic_try_cmpxchg(safe, unsafe, 42)); > > + > > KUNIT_EXPECT_KASAN_FAIL(test, atomic_sub_and_test(42, unsafe)); > > KUNIT_EXPECT_KASAN_FAIL(test, atomic_dec_and_test(unsafe)); > > KUNIT_EXPECT_KASAN_FAIL(test, atomic_inc_and_test(unsafe)); > > KUNIT_EXPECT_KASAN_FAIL(test, atomic_add_negative(42, unsafe)); > > - KUNIT_EXPECT_KASAN_FAIL(test, atomic_add_unless(unsafe, 21, 42)); > > - KUNIT_EXPECT_KASAN_FAIL(test, atomic_inc_not_zero(unsafe)); > > - KUNIT_EXPECT_KASAN_FAIL(test, atomic_inc_unless_negative(unsafe)); > > - KUNIT_EXPECT_KASAN_FAIL(test, atomic_dec_unless_positive(unsafe)); > > - KUNIT_EXPECT_KASAN_FAIL(test, atomic_dec_if_positive(unsafe)); > > > > - KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_read(unsafe)); > > + /* > > + * The result of the test below may vary due to garbage values of unsafe in > > + * store-only mode. Therefore, skip this test when KASAN is configured > > + * in store-only mode. > > + */ > > + if (!kasan_stonly_enabled()) { > > + KUNIT_EXPECT_KASAN_FAIL(test, atomic_add_unless(unsafe, 21, 42)); > > + KUNIT_EXPECT_KASAN_FAIL(test, atomic_inc_not_zero(unsafe)); > > + KUNIT_EXPECT_KASAN_FAIL(test, atomic_inc_unless_negative(unsafe)); > > + KUNIT_EXPECT_KASAN_FAIL(test, atomic_dec_unless_positive(unsafe)); > > + KUNIT_EXPECT_KASAN_FAIL(test, atomic_dec_if_positive(unsafe)); > > + } > > + > > + if (kasan_stonly_enabled()) > > + KUNIT_EXPECT_KASAN_SUCCESS(test, atomic_long_read(unsafe)); > > + else > > + KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_read(unsafe)); > > + > > KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_set(unsafe, 42)); > > KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_add(42, unsafe)); > > KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_sub(42, unsafe)); > > @@ -776,16 +905,32 @@ static void kasan_atomics_helper(struct kunit *test, void *unsafe, void *safe) > > KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_xchg(unsafe, 42)); > > KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_cmpxchg(unsafe, 21, 42)); > > KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_try_cmpxchg(unsafe, safe, 42)); > > - KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_try_cmpxchg(safe, unsafe, 42)); > > + > > + /* > > + * The result of the test below may vary due to garbage values in > > + * store-only mode. Therefore, skip this test when KASAN is configured > > + * in store-only mode. > > + */ > > + if (!kasan_stonly_enabled()) > > + KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_try_cmpxchg(safe, unsafe, 42)); > > + > > KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_sub_and_test(42, unsafe)); > > KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_dec_and_test(unsafe)); > > KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_inc_and_test(unsafe)); > > KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_add_negative(42, unsafe)); > > - KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_add_unless(unsafe, 21, 42)); > > - KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_inc_not_zero(unsafe)); > > - KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_inc_unless_negative(unsafe)); > > - KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_dec_unless_positive(unsafe)); > > - KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_dec_if_positive(unsafe)); > > + > > + /* > > + * The result of the test below may vary due to garbage values in > > + * store-only mode. Therefore, skip this test when KASAN is configured > > + * in store-only mode. > > + */ > > + if (!kasan_stonly_enabled()) { > > + KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_add_unless(unsafe, 21, 42)); > > + KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_inc_not_zero(unsafe)); > > + KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_inc_unless_negative(unsafe)); > > + KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_dec_unless_positive(unsafe)); > > + KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_dec_if_positive(unsafe)); > > + } > > } > > > > static void kasan_atomics(struct kunit *test) > > @@ -842,8 +987,18 @@ static void ksize_unpoisons_memory(struct kunit *test) > > /* These must trigger a KASAN report. */ > > if (IS_ENABLED(CONFIG_KASAN_GENERIC)) > > KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[size]); > > - KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[size + 5]); > > - KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[real_size - 1]); > > + > > + if (kasan_stonly_enabled()) { > > + KUNIT_EXPECT_KASAN_SUCCESS(test, ((volatile char *)ptr)[size + 5]); > > + KUNIT_EXPECT_KASAN_SUCCESS(test, ((volatile char *)ptr)[real_size - 1]); > > + if (!kasan_sync_fault_possible()) { > > + KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[size + 5] = 0); > > + KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[real_size - 1] = 0); > > + } > > + } else { > > + KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[size + 5]); > > + KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[real_size - 1]); > > + } > > > > kfree(ptr); > > } > > @@ -863,8 +1018,17 @@ static void ksize_uaf(struct kunit *test) > > > > OPTIMIZER_HIDE_VAR(ptr); > > KUNIT_EXPECT_KASAN_FAIL(test, ksize(ptr)); > > - KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[0]); > > - KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[size]); > > + if (kasan_stonly_enabled()) { > > + KUNIT_EXPECT_KASAN_SUCCESS(test, ((volatile char *)ptr)[0]); > > + KUNIT_EXPECT_KASAN_SUCCESS(test, ((volatile char *)ptr)[size]); > > + if (!kasan_sync_fault_possible()) { > > + KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[0] = 0); > > + KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[size] = 0); > > + } > > + } else { > > + KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[0]); > > + KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[size]); > > + } > > } > > > > /* > > @@ -886,7 +1050,11 @@ static void rcu_uaf_reclaim(struct rcu_head *rp) > > container_of(rp, struct kasan_rcu_info, rcu); > > > > kfree(fp); > > - ((volatile struct kasan_rcu_info *)fp)->i; > > + > > + if (kasan_stonly_enabled() && !kasan_async_fault_possible()) > > + ((volatile struct kasan_rcu_info *)fp)->i = 0; > > + else > > + ((volatile struct kasan_rcu_info *)fp)->i; > > } > > > > static void rcu_uaf(struct kunit *test) > > @@ -899,9 +1067,14 @@ static void rcu_uaf(struct kunit *test) > > global_rcu_ptr = rcu_dereference_protected( > > (struct kasan_rcu_info __rcu *)ptr, NULL); > > > > - KUNIT_EXPECT_KASAN_FAIL(test, > > - call_rcu(&global_rcu_ptr->rcu, rcu_uaf_reclaim); > > - rcu_barrier()); > > + if (kasan_stonly_enabled() && kasan_async_fault_possible()) > > + KUNIT_EXPECT_KASAN_SUCCESS(test, > > + call_rcu(&global_rcu_ptr->rcu, rcu_uaf_reclaim); > > + rcu_barrier()); > > + else > > + KUNIT_EXPECT_KASAN_FAIL(test, > > + call_rcu(&global_rcu_ptr->rcu, rcu_uaf_reclaim); > > + rcu_barrier()); > > } > > > > static void workqueue_uaf_work(struct work_struct *work) > > @@ -924,8 +1097,12 @@ static void workqueue_uaf(struct kunit *test) > > queue_work(workqueue, work); > > destroy_workqueue(workqueue); > > > > - KUNIT_EXPECT_KASAN_FAIL(test, > > - ((volatile struct work_struct *)work)->data); > > + if (kasan_stonly_enabled()) > > + KUNIT_EXPECT_KASAN_SUCCESS(test, > > + ((volatile struct work_struct *)work)->data); > > + else > > + KUNIT_EXPECT_KASAN_FAIL(test, > > + ((volatile struct work_struct *)work)->data); > > } > > > > static void kfree_via_page(struct kunit *test) > > @@ -972,7 +1149,12 @@ static void kmem_cache_oob(struct kunit *test) > > return; > > } > > > > - KUNIT_EXPECT_KASAN_FAIL(test, *p = p[size + OOB_TAG_OFF]); > > + if (kasan_stonly_enabled()) { > > + KUNIT_EXPECT_KASAN_SUCCESS(test, *p = p[size + OOB_TAG_OFF]); > > + if (!kasan_async_fault_possible()) > > + KUNIT_EXPECT_KASAN_FAIL(test, p[size + OOB_TAG_OFF] = *p); > > + } else > > + KUNIT_EXPECT_KASAN_FAIL(test, *p = p[size + OOB_TAG_OFF]); > > > > kmem_cache_free(cache, p); > > kmem_cache_destroy(cache); > > @@ -1068,7 +1250,12 @@ static void kmem_cache_rcu_uaf(struct kunit *test) > > */ > > rcu_barrier(); > > > > - KUNIT_EXPECT_KASAN_FAIL(test, READ_ONCE(*p)); > > + if (kasan_stonly_enabled()) { > > + KUNIT_EXPECT_KASAN_SUCCESS(test, READ_ONCE(*p)); > > + if (!kasan_async_fault_possible()) > > + KUNIT_EXPECT_KASAN_FAIL(test, WRITE_ONCE(*p, 0)); > > + } else > > + KUNIT_EXPECT_KASAN_FAIL(test, READ_ONCE(*p)); > > > > kmem_cache_destroy(cache); > > } > > @@ -1206,7 +1393,13 @@ static void mempool_oob_right_helper(struct kunit *test, mempool_t *pool, size_t > > if (IS_ENABLED(CONFIG_KASAN_GENERIC)) > > KUNIT_EXPECT_KASAN_FAIL(test, > > ((volatile char *)&elem[size])[0]); > > - else > > + else if (kasan_stonly_enabled()) { > > + KUNIT_EXPECT_KASAN_SUCCESS(test, > > + ((volatile char *)&elem[round_up(size, KASAN_GRANULE_SIZE)])[0]); > > + if (!kasan_async_fault_possible()) > > + KUNIT_EXPECT_KASAN_FAIL(test, > > + ((volatile char *)&elem[round_up(size, KASAN_GRANULE_SIZE)])[0] = 0); > > + } else > > KUNIT_EXPECT_KASAN_FAIL(test, > > ((volatile char *)&elem[round_up(size, KASAN_GRANULE_SIZE)])[0]); > > > > @@ -1273,7 +1466,13 @@ static void mempool_uaf_helper(struct kunit *test, mempool_t *pool, bool page) > > mempool_free(elem, pool); > > > > ptr = page ? page_address((struct page *)elem) : elem; > > - KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[0]); > > + > > + if (kasan_stonly_enabled()) { > > + KUNIT_EXPECT_KASAN_SUCCESS(test, ((volatile char *)ptr)[0]); > > + if (!kasan_async_fault_possible()) > > + KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[0] = 0); > > + } else > > + KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[0]); > > } > > > > static void mempool_kmalloc_uaf(struct kunit *test) > > @@ -1532,8 +1731,13 @@ static void kasan_memchr(struct kunit *test) > > > > OPTIMIZER_HIDE_VAR(ptr); > > OPTIMIZER_HIDE_VAR(size); > > - KUNIT_EXPECT_KASAN_FAIL(test, > > - kasan_ptr_result = memchr(ptr, '1', size + 1)); > > + > > + if (kasan_stonly_enabled()) > > + KUNIT_EXPECT_KASAN_SUCCESS(test, > > + kasan_ptr_result = memchr(ptr, '1', size + 1)); > > + else > > + KUNIT_EXPECT_KASAN_FAIL(test, > > + kasan_ptr_result = memchr(ptr, '1', size + 1)); > > > > kfree(ptr); > > } > > @@ -1559,8 +1763,14 @@ static void kasan_memcmp(struct kunit *test) > > > > OPTIMIZER_HIDE_VAR(ptr); > > OPTIMIZER_HIDE_VAR(size); > > - KUNIT_EXPECT_KASAN_FAIL(test, > > - kasan_int_result = memcmp(ptr, arr, size+1)); > > + > > + if (kasan_stonly_enabled()) > > + KUNIT_EXPECT_KASAN_SUCCESS(test, > > + kasan_int_result = memcmp(ptr, arr, size+1)); > > + else > > + KUNIT_EXPECT_KASAN_FAIL(test, > > + kasan_int_result = memcmp(ptr, arr, size+1)); > > + > > kfree(ptr); > > } > > > > @@ -1593,9 +1803,16 @@ static void kasan_strings(struct kunit *test) > > KUNIT_EXPECT_EQ(test, KASAN_GRANULE_SIZE - 2, > > strscpy(ptr, src + 1, KASAN_GRANULE_SIZE)); > > > > - /* strscpy should fail if the first byte is unreadable. */ > > - KUNIT_EXPECT_KASAN_FAIL(test, strscpy(ptr, src + KASAN_GRANULE_SIZE, > > - KASAN_GRANULE_SIZE)); > > + if (kasan_stonly_enabled()) { > > + KUNIT_EXPECT_KASAN_SUCCESS(test, strscpy(ptr, src + KASAN_GRANULE_SIZE, > > + KASAN_GRANULE_SIZE)); > > + if (!kasan_async_fault_possible()) > > + /* strscpy should fail when the first byte is to be written. */ > > + KUNIT_EXPECT_KASAN_FAIL(test, strscpy(ptr + size, src, KASAN_GRANULE_SIZE)); > > + } else > > + /* strscpy should fail if the first byte is unreadable. */ > > + KUNIT_EXPECT_KASAN_FAIL(test, strscpy(ptr, src + KASAN_GRANULE_SIZE, > > + KASAN_GRANULE_SIZE)); > > > > kfree(src); > > kfree(ptr); > > @@ -1607,17 +1824,22 @@ static void kasan_strings(struct kunit *test) > > * will likely point to zeroed byte. > > */ > > ptr += 16; > > - KUNIT_EXPECT_KASAN_FAIL(test, kasan_ptr_result = strchr(ptr, '1')); > > > > - KUNIT_EXPECT_KASAN_FAIL(test, kasan_ptr_result = strrchr(ptr, '1')); > > - > > - KUNIT_EXPECT_KASAN_FAIL(test, kasan_int_result = strcmp(ptr, "2")); > > - > > - KUNIT_EXPECT_KASAN_FAIL(test, kasan_int_result = strncmp(ptr, "2", 1)); > > - > > - KUNIT_EXPECT_KASAN_FAIL(test, kasan_int_result = strlen(ptr)); > > - > > - KUNIT_EXPECT_KASAN_FAIL(test, kasan_int_result = strnlen(ptr, 1)); > > + if (kasan_stonly_enabled()) { > > + KUNIT_EXPECT_KASAN_SUCCESS(test, kasan_ptr_result = strchr(ptr, '1')); > > + KUNIT_EXPECT_KASAN_SUCCESS(test, kasan_ptr_result = strrchr(ptr, '1')); > > + KUNIT_EXPECT_KASAN_SUCCESS(test, kasan_int_result = strcmp(ptr, "2")); > > + KUNIT_EXPECT_KASAN_SUCCESS(test, kasan_int_result = strncmp(ptr, "2", 1)); > > + KUNIT_EXPECT_KASAN_SUCCESS(test, kasan_int_result = strlen(ptr)); > > + KUNIT_EXPECT_KASAN_SUCCESS(test, kasan_int_result = strnlen(ptr, 1)); > > + } else { > > + KUNIT_EXPECT_KASAN_FAIL(test, kasan_ptr_result = strchr(ptr, '1')); > > + KUNIT_EXPECT_KASAN_FAIL(test, kasan_ptr_result = strrchr(ptr, '1')); > > + KUNIT_EXPECT_KASAN_FAIL(test, kasan_int_result = strcmp(ptr, "2")); > > + KUNIT_EXPECT_KASAN_FAIL(test, kasan_int_result = strncmp(ptr, "2", 1)); > > + KUNIT_EXPECT_KASAN_FAIL(test, kasan_int_result = strlen(ptr)); > > + KUNIT_EXPECT_KASAN_FAIL(test, kasan_int_result = strnlen(ptr, 1)); > > + } > > } > > > > static void kasan_bitops_modify(struct kunit *test, int nr, void *addr) > > @@ -1636,12 +1858,27 @@ static void kasan_bitops_test_and_modify(struct kunit *test, int nr, void *addr) > > { > > KUNIT_EXPECT_KASAN_FAIL(test, test_and_set_bit(nr, addr)); > > KUNIT_EXPECT_KASAN_FAIL(test, __test_and_set_bit(nr, addr)); > > - KUNIT_EXPECT_KASAN_FAIL(test, test_and_set_bit_lock(nr, addr)); > > + > > + /* > > + * When KASAN is running in store-only mode, > > + * a fault won't occur even if the bit is set. > > + * Therefore, skip the test_and_set_bit_lock test in store-only mode. > > + */ > > + if (!kasan_stonly_enabled()) > > + KUNIT_EXPECT_KASAN_FAIL(test, test_and_set_bit_lock(nr, addr)); > > + > > KUNIT_EXPECT_KASAN_FAIL(test, test_and_clear_bit(nr, addr)); > > KUNIT_EXPECT_KASAN_FAIL(test, __test_and_clear_bit(nr, addr)); > > KUNIT_EXPECT_KASAN_FAIL(test, test_and_change_bit(nr, addr)); > > KUNIT_EXPECT_KASAN_FAIL(test, __test_and_change_bit(nr, addr)); > > - KUNIT_EXPECT_KASAN_FAIL(test, kasan_int_result = test_bit(nr, addr)); > > + > > + if (kasan_stonly_enabled()) { > > + KUNIT_EXPECT_KASAN_SUCCESS(test, kasan_int_result = test_bit(nr, addr)); > > + if (!kasan_async_fault_possible()) > > + KUNIT_EXPECT_KASAN_FAIL(test, set_bit(nr, addr)); > > + } else > > + KUNIT_EXPECT_KASAN_FAIL(test, kasan_int_result = test_bit(nr, addr)); > > + > > if (nr < 7) > > KUNIT_EXPECT_KASAN_FAIL(test, kasan_int_result = > > xor_unlock_is_negative_byte(1 << nr, addr)); > > @@ -1765,7 +2002,12 @@ static void vmalloc_oob(struct kunit *test) > > KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)v_ptr)[size]); > > > > /* An aligned access into the first out-of-bounds granule. */ > > - KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)v_ptr)[size + 5]); > > + if (kasan_stonly_enabled()) { > > + KUNIT_EXPECT_KASAN_SUCCESS(test, ((volatile char *)v_ptr)[size + 5]); > > + if (!kasan_async_fault_possible()) > > + KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)v_ptr)[size + 5] = 0); > > + } else > > + KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)v_ptr)[size + 5]); > > > > /* Check that in-bounds accesses to the physical page are valid. */ > > page = vmalloc_to_page(v_ptr); > > @@ -2042,16 +2284,33 @@ static void copy_user_test_oob(struct kunit *test) > > > > KUNIT_EXPECT_KASAN_FAIL(test, > > unused = copy_from_user(kmem, usermem, size + 1)); > > - KUNIT_EXPECT_KASAN_FAIL(test, > > - unused = copy_to_user(usermem, kmem, size + 1)); > > + > > + if (kasan_stonly_enabled()) > > + KUNIT_EXPECT_KASAN_SUCCESS(test, > > + unused = copy_to_user(usermem, kmem, size + 1)); > > + else > > + KUNIT_EXPECT_KASAN_FAIL(test, > > + unused = copy_to_user(usermem, kmem, size + 1)); > > + > > KUNIT_EXPECT_KASAN_FAIL(test, > > unused = __copy_from_user(kmem, usermem, size + 1)); > > - KUNIT_EXPECT_KASAN_FAIL(test, > > - unused = __copy_to_user(usermem, kmem, size + 1)); > > + > > + if (kasan_stonly_enabled()) > > + KUNIT_EXPECT_KASAN_SUCCESS(test, > > + unused = __copy_to_user(usermem, kmem, size + 1)); > > + else > > + KUNIT_EXPECT_KASAN_FAIL(test, > > + unused = __copy_to_user(usermem, kmem, size + 1)); > > + > > KUNIT_EXPECT_KASAN_FAIL(test, > > unused = __copy_from_user_inatomic(kmem, usermem, size + 1)); > > - KUNIT_EXPECT_KASAN_FAIL(test, > > - unused = __copy_to_user_inatomic(usermem, kmem, size + 1)); > > + > > + if (kasan_stonly_enabled()) > > + KUNIT_EXPECT_KASAN_SUCCESS(test, > > + unused = __copy_to_user_inatomic(usermem, kmem, size + 1)); > > + else > > + KUNIT_EXPECT_KASAN_FAIL(test, > > + unused = __copy_to_user_inatomic(usermem, kmem, size + 1)); > > > > /* > > * Prepare a long string in usermem to avoid the strncpy_from_user test > > -- > > LEVI:{C3F47F37-75D8-414A-A8BA-3980EC8A46D7} > > > > This patch does not look good. > > Right now, KASAN tests are crafted to avoid/self-contain harmful > memory corruptions that they do (e.g. make sure that OOB write > accesses land in in-object kmalloc training space, etc.). If you turn > read accesses in tests into write accesses, memory corruptions caused > by the earlier tests will crash the kernel or the latter tests. That's why I run the store-only test when this mode is "sync" In case of "async/asymm" as you mention since it reports "after", there will be memory corruption. But in case of sync, when the MTE fault happens, it doesn't write to memory so, I think it's fine. > > The easiest thing to do for now is to disable the tests that check bad > read accesses when store-only is enabled. > > If we want to convert tests into doing write accesses instead of > reads, this needs to be done separately for each test (i.e. via a > separate patch) with an explanation why doing this is safe (and > adjustments whenever it's not). And we need a better way to code this > instead of the horrifying number of if/else checks. > > Thank you! Hmm, as I mention above, the testcase with store-only/sync mode seems to fine. But, If the "testcase" is failed, as you mention it makes a memory corruption. If success case is fine, Please let me make all related story-only case be seperated to each function (but almost simliar to pre-exist testcase) with sync mode otherwise, let me seperate them just checking it whether it success when it accesses to invalid memory with read/fetch. Thanks :) -- Sincerely, Yeoreum Yun