Lines Matching +full:0 +full:x61

44 		"	sllg 1,1,12(0)\n"  in guest_do_one_essa()
46 " .insn rrf,0xb9ab0000,2,1,1,0\n" in guest_do_one_essa()
48 " diag 0,0,0x501\n" in guest_do_one_essa()
49 "0: j 0b" in guest_do_one_essa()
71 "1: sllg 2,1,12(0)\n" in guest_dirty_test_data()
73 " .insn rrf,0xb9ab0000,4,2,1,0\n" in guest_dirty_test_data()
79 " diag 0,0,0x501\n" in guest_dirty_test_data()
80 "0: j 0b" in guest_dirty_test_data()
106 vm_userspace_mem_region_add(vm, VM_MEM_SRC_ANONYMOUS, 0, 0, MAIN_PAGE_COUNT, 0); in create_main_memslot()
108 for (i = 0; i < NR_MEM_REGIONS; i++) in create_main_memslot()
109 vm->memslots[i] = 0; in create_main_memslot()
119 0 in create_test_memslot()
131 * | 0 | | in create_memslots()
150 slot0 = memslot2region(vm, 0); in finish_vm_setup()
179 vm_mem_region_set_flags(vm, 0, KVM_MEM_LOG_DIRTY_PAGES); in enable_dirty_tracking()
218 errno = 0; in vm_get_cmma_bits()
221 .start_gfn = 0, in vm_get_cmma_bits()
224 .values = (__u64)&cmma_value_buf[0] in vm_get_cmma_bits()
239 rc = vm_get_cmma_bits(vm, 0, &errno_out); in test_get_cmma_basic()
249 rc = vm_get_cmma_bits(vm, 0, &errno_out); in test_get_cmma_basic()
255 TEST_ASSERT_EQ(rc, 0); in test_get_cmma_basic()
256 TEST_ASSERT_EQ(errno_out, 0); in test_get_cmma_basic()
262 rc = vm_get_cmma_bits(vm, 0xfeedc0fe, &errno_out); in test_get_cmma_basic()
273 TEST_ASSERT_EQ(vcpu->run->s390_sieic.ipa, 0x8300); in assert_exit_was_hypercall()
274 TEST_ASSERT_EQ(vcpu->run->s390_sieic.ipb, 0x5010000); in assert_exit_was_hypercall()
289 errno = 0; in test_migration_mode()
310 errno = 0; in test_migration_mode()
317 TEST_ASSERT_EQ(rc, 0); in test_migration_mode()
319 errno = 0; in test_migration_mode()
336 0 in test_migration_mode()
353 TEST_ASSERT_EQ(rc, 0); in test_migration_mode()
355 errno = 0; in test_migration_mode()
362 vm_mem_region_set_flags(vm, TEST_DATA_TWO_MEMSLOT, 0); in test_migration_mode()
390 .start_gfn = 0, in assert_all_slots_cmma_dirty()
392 .flags = 0, in assert_all_slots_cmma_dirty()
393 .values = (__u64)&cmma_value_buf[0] in assert_all_slots_cmma_dirty()
395 memset(cmma_value_buf, 0xff, sizeof(cmma_value_buf)); in assert_all_slots_cmma_dirty()
399 TEST_ASSERT_EQ(args.start_gfn, 0); in assert_all_slots_cmma_dirty()
405 .flags = 0, in assert_all_slots_cmma_dirty()
406 .values = (__u64)&cmma_value_buf[0] in assert_all_slots_cmma_dirty()
408 memset(cmma_value_buf, 0xff, sizeof(cmma_value_buf)); in assert_all_slots_cmma_dirty()
412 TEST_ASSERT_EQ(args.remaining, 0); in assert_all_slots_cmma_dirty()
418 .flags = 0, in assert_all_slots_cmma_dirty()
419 .values = (__u64)&cmma_value_buf[0] in assert_all_slots_cmma_dirty()
421 memset(cmma_value_buf, 0xff, sizeof(cmma_value_buf)); in assert_all_slots_cmma_dirty()
423 TEST_ASSERT_EQ(args.count, 0); in assert_all_slots_cmma_dirty()
424 TEST_ASSERT_EQ(args.start_gfn, 0); in assert_all_slots_cmma_dirty()
425 TEST_ASSERT_EQ(args.remaining, 0); in assert_all_slots_cmma_dirty()
435 /* If we start from GFN 0 again, nothing should be dirty. */ in assert_no_pages_cmma_dirty()
437 .start_gfn = 0, in assert_no_pages_cmma_dirty()
439 .flags = 0, in assert_no_pages_cmma_dirty()
440 .values = (__u64)&cmma_value_buf[0] in assert_no_pages_cmma_dirty()
442 memset(cmma_value_buf, 0xff, sizeof(cmma_value_buf)); in assert_no_pages_cmma_dirty()
445 TEST_FAIL("pages are still dirty start_gfn=0x%llx count=%u remaining=%llu", in assert_no_pages_cmma_dirty()
485 .flags = 0, in query_cmma_range()
486 .values = (__u64)&cmma_value_buf[0] in query_cmma_range()
488 memset(cmma_value_buf, 0xff, sizeof(cmma_value_buf)); in query_cmma_range()
503 for (size_t i = 0; i < dirty_gfn_count; i++) in assert_cmma_dirty()
504 TEST_ASSERT_EQ(cmma_value_buf[0], 0x0); /* stable state */ in assert_cmma_dirty()
505 TEST_ASSERT_EQ(cmma_value_buf[dirty_gfn_count], 0xff); /* not touched */ in assert_cmma_dirty()
540 * Query CMMA attributes of one page, starting at page 0. Since the in test_get_skip_holes()
544 * 0: not dirty in test_get_skip_holes()
545 * [0x1, 0x200): dirty in test_get_skip_holes()
547 query_cmma_range(vm, 0, 1, &log); in test_get_skip_holes()
552 * Query CMMA attributes of 32 (0x20) pages past the end of the TEST_DATA in test_get_skip_holes()
556 * [0, 0x21): not dirty in test_get_skip_holes()
557 * [0x21, 0x200): dirty in test_get_skip_holes()
559 query_cmma_range(vm, TEST_DATA_START_GFN + TEST_DATA_PAGE_COUNT, 0x20, &log); in test_get_skip_holes()
560 assert_cmma_dirty(gfn_offset, 0x20, &log); in test_get_skip_holes()
561 gfn_offset += 0x20; in test_get_skip_holes()
564 gfn_offset += 0x20; in test_get_skip_holes()
567 * After skipping 32 pages, query the next 32 (0x20) pages. in test_get_skip_holes()
569 * [0, 0x21): not dirty in test_get_skip_holes()
570 * [0x21, 0x41): dirty in test_get_skip_holes()
571 * [0x41, 0x61): not dirty in test_get_skip_holes()
572 * [0x61, 0x200): dirty in test_get_skip_holes()
574 query_cmma_range(vm, gfn_offset, 0x20, &log); in test_get_skip_holes()
575 assert_cmma_dirty(gfn_offset, 0x20, &log); in test_get_skip_holes()
576 gfn_offset += 0x20; in test_get_skip_holes()
580 * yield page 0x21. in test_get_skip_holes()
582 * [0, 0x22): not dirty in test_get_skip_holes()
583 * [0x22, 0x41): dirty in test_get_skip_holes()
584 * [0x41, 0x61): not dirty in test_get_skip_holes()
585 * [0x61, 0x200): dirty in test_get_skip_holes()
588 assert_cmma_dirty(TEST_DATA_START_GFN + 0x21, 1, &log); in test_get_skip_holes()
592 * Query 15 (0xF) pages from page 0x23 in TEST_DATA memslot. in test_get_skip_holes()
593 * This should yield pages [0x23, 0x33). in test_get_skip_holes()
595 * [0, 0x22): not dirty in test_get_skip_holes()
596 * 0x22: dirty in test_get_skip_holes()
597 * [0x23, 0x33): not dirty in test_get_skip_holes()
598 * [0x33, 0x41): dirty in test_get_skip_holes()
599 * [0x41, 0x61): not dirty in test_get_skip_holes()
600 * [0x61, 0x200): dirty in test_get_skip_holes()
602 gfn_offset = TEST_DATA_START_GFN + 0x23; in test_get_skip_holes()
607 * Query 17 (0x11) pages from page 0x22 in TEST_DATA memslot. in test_get_skip_holes()
608 * This should yield page [0x22, 0x33) in test_get_skip_holes()
610 * [0, 0x33): not dirty in test_get_skip_holes()
611 * [0x33, 0x41): dirty in test_get_skip_holes()
612 * [0x41, 0x61): not dirty in test_get_skip_holes()
613 * [0x61, 0x200): dirty in test_get_skip_holes()
615 gfn_offset = TEST_DATA_START_GFN + 0x22; in test_get_skip_holes()
620 * Query 25 (0x19) pages from page 0x40 in TEST_DATA memslot. in test_get_skip_holes()
621 * This should yield page 0x40 and nothing more, since there are more in test_get_skip_holes()
622 * than 16 non-dirty pages after page 0x40. in test_get_skip_holes()
624 * [0, 0x33): not dirty in test_get_skip_holes()
625 * [0x33, 0x40): dirty in test_get_skip_holes()
626 * [0x40, 0x61): not dirty in test_get_skip_holes()
627 * [0x61, 0x200): dirty in test_get_skip_holes()
629 gfn_offset = TEST_DATA_START_GFN + 0x40; in test_get_skip_holes()
634 * Query pages [0x33, 0x40). in test_get_skip_holes()
636 * [0, 0x61): not dirty in test_get_skip_holes()
637 * [0x61, 0x200): dirty in test_get_skip_holes()
639 gfn_offset = TEST_DATA_START_GFN + 0x33; in test_get_skip_holes()
640 query_cmma_range(vm, gfn_offset, 0x40 - 0x33, &log); in test_get_skip_holes()
641 assert_cmma_dirty(gfn_offset, 0x40 - 0x33, &log); in test_get_skip_holes()
644 * Query the remaining pages [0x61, 0x200). in test_get_skip_holes()
647 query_cmma_range(vm, gfn_offset, TEST_DATA_PAGE_COUNT - 0x61, &log); in test_get_skip_holes()
648 assert_cmma_dirty(TEST_DATA_START_GFN + 0x61, TEST_DATA_PAGE_COUNT - 0x61, &log); in test_get_skip_holes()
694 for (idx = 0; idx < ARRAY_SIZE(testlist); idx++) { in main()