Lines Matching refs:iclog

34 	struct xlog_in_core	**iclog,
50 struct xlog_in_core *iclog);
56 STATIC void xlog_state_done_syncing(xlog_in_core_t *iclog, int);
61 struct xlog_in_core *iclog);
66 struct xlog_in_core **iclog,
73 struct xlog_in_core *iclog);
77 struct xlog_in_core *iclog,
82 struct xlog_in_core *iclog);
108 struct xlog_in_core *iclog,
114 struct xlog_in_core *iclog,
502 struct xlog_in_core **iclog, in xfs_log_done() argument
514 (xlog_commit_record(log, ticket, iclog, &lsn)))) { in xfs_log_done()
551 struct xlog_in_core *iclog, in xfs_log_notify() argument
556 spin_lock(&iclog->ic_callback_lock); in xfs_log_notify()
557 abortflg = (iclog->ic_state & XLOG_STATE_IOERROR); in xfs_log_notify()
559 ASSERT_ALWAYS((iclog->ic_state == XLOG_STATE_ACTIVE) || in xfs_log_notify()
560 (iclog->ic_state == XLOG_STATE_WANT_SYNC)); in xfs_log_notify()
562 *(iclog->ic_callback_tail) = cb; in xfs_log_notify()
563 iclog->ic_callback_tail = &(cb->cb_next); in xfs_log_notify()
565 spin_unlock(&iclog->ic_callback_lock); in xfs_log_notify()
572 struct xlog_in_core *iclog) in xfs_log_release_iclog() argument
574 if (xlog_state_release_iclog(mp->m_log, iclog)) { in xfs_log_release_iclog()
848 struct xlog_in_core *iclog; in xfs_log_write_unmount_record() local
884 iclog = log->l_iclog; in xfs_log_write_unmount_record()
885 atomic_inc(&iclog->ic_refcnt); in xfs_log_write_unmount_record()
886 xlog_state_want_sync(log, iclog); in xfs_log_write_unmount_record()
888 error = xlog_state_release_iclog(log, iclog); in xfs_log_write_unmount_record()
891 switch (iclog->ic_state) { in xfs_log_write_unmount_record()
894 xlog_wait(&iclog->ic_force_wait, &log->l_icloglock); in xfs_log_write_unmount_record()
923 xlog_in_core_t *iclog; in xfs_log_unmount_write() local
943 first_iclog = iclog = log->l_iclog; in xfs_log_unmount_write()
945 if (!(iclog->ic_state & XLOG_STATE_IOERROR)) { in xfs_log_unmount_write()
946 ASSERT(iclog->ic_state & XLOG_STATE_ACTIVE); in xfs_log_unmount_write()
947 ASSERT(iclog->ic_offset == 0); in xfs_log_unmount_write()
949 iclog = iclog->ic_next; in xfs_log_unmount_write()
950 } while (iclog != first_iclog); in xfs_log_unmount_write()
969 iclog = log->l_iclog; in xfs_log_unmount_write()
970 atomic_inc(&iclog->ic_refcnt); in xfs_log_unmount_write()
972 xlog_state_want_sync(log, iclog); in xfs_log_unmount_write()
974 error = xlog_state_release_iclog(log, iclog); in xfs_log_unmount_write()
978 if ( ! ( iclog->ic_state == XLOG_STATE_ACTIVE in xfs_log_unmount_write()
979 || iclog->ic_state == XLOG_STATE_DIRTY in xfs_log_unmount_write()
980 || iclog->ic_state == XLOG_STATE_IOERROR) ) { in xfs_log_unmount_write()
982 xlog_wait(&iclog->ic_force_wait, in xfs_log_unmount_write()
1255 struct xlog_in_core *iclog = bp->b_log_item; in xlog_iodone() local
1256 struct xlog *l = iclog->ic_log; in xlog_iodone()
1265 iclog->ic_state & XLOG_STATE_IOABORT) { in xlog_iodone()
1266 if (iclog->ic_state & XLOG_STATE_IOABORT) in xlog_iodone()
1267 iclog->ic_state &= ~XLOG_STATE_IOABORT; in xlog_iodone()
1278 } else if (iclog->ic_state & XLOG_STATE_IOERROR) { in xlog_iodone()
1284 xlog_state_done_syncing(iclog, aborted); in xlog_iodone()
1423 xlog_in_core_t *iclog, *prev_iclog=NULL; in xlog_alloc_log() local
1523 iclog = *iclogp; in xlog_alloc_log()
1524 iclog->ic_prev = prev_iclog; in xlog_alloc_log()
1525 prev_iclog = iclog; in xlog_alloc_log()
1539 iclog->ic_bp = bp; in xlog_alloc_log()
1540 iclog->ic_data = bp->b_addr; in xlog_alloc_log()
1542 log->l_iclog_bak[i] = &iclog->ic_header; in xlog_alloc_log()
1544 head = &iclog->ic_header; in xlog_alloc_log()
1554 iclog->ic_size = BBTOB(bp->b_length) - log->l_iclog_hsize; in xlog_alloc_log()
1555 iclog->ic_state = XLOG_STATE_ACTIVE; in xlog_alloc_log()
1556 iclog->ic_log = log; in xlog_alloc_log()
1557 atomic_set(&iclog->ic_refcnt, 0); in xlog_alloc_log()
1558 spin_lock_init(&iclog->ic_callback_lock); in xlog_alloc_log()
1559 iclog->ic_callback_tail = &(iclog->ic_callback); in xlog_alloc_log()
1560 iclog->ic_datap = (char *)iclog->ic_data + log->l_iclog_hsize; in xlog_alloc_log()
1562 init_waitqueue_head(&iclog->ic_force_wait); in xlog_alloc_log()
1563 init_waitqueue_head(&iclog->ic_write_wait); in xlog_alloc_log()
1565 iclogp = &iclog->ic_next; in xlog_alloc_log()
1576 for (iclog = log->l_iclog; iclog; iclog = prev_iclog) { in xlog_alloc_log()
1577 prev_iclog = iclog->ic_next; in xlog_alloc_log()
1578 if (iclog->ic_bp) in xlog_alloc_log()
1579 xfs_buf_free(iclog->ic_bp); in xlog_alloc_log()
1580 kmem_free(iclog); in xlog_alloc_log()
1599 struct xlog_in_core **iclog, in xlog_commit_record() argument
1614 ASSERT_ALWAYS(iclog); in xlog_commit_record()
1615 error = xlog_write(log, &vec, ticket, commitlsnp, iclog, in xlog_commit_record()
1691 struct xlog_in_core *iclog, in xlog_pack_data() argument
1695 int size = iclog->ic_offset + roundoff; in xlog_pack_data()
1699 cycle_lsn = CYCLE_LSN_DISK(iclog->ic_header.h_lsn); in xlog_pack_data()
1701 dp = iclog->ic_datap; in xlog_pack_data()
1705 iclog->ic_header.h_cycle_data[i] = *(__be32 *)dp; in xlog_pack_data()
1711 xlog_in_core_2_t *xhdr = iclog->ic_data; in xlog_pack_data()
1786 struct xlog_in_core *iclog = bp->b_log_item; in xlog_bdstrat() local
1789 if (iclog->ic_state & XLOG_STATE_IOERROR) { in xlog_bdstrat()
1834 struct xlog_in_core *iclog) in xlog_sync() argument
1847 ASSERT(atomic_read(&iclog->ic_refcnt) == 0); in xlog_sync()
1850 count_init = log->l_iclog_hsize + iclog->ic_offset; in xlog_sync()
1872 xlog_pack_data(log, iclog, roundoff); in xlog_sync()
1875 size = iclog->ic_offset; in xlog_sync()
1878 iclog->ic_header.h_len = cpu_to_be32(size); in xlog_sync()
1880 bp = iclog->ic_bp; in xlog_sync()
1881 XFS_BUF_SET_ADDR(bp, BLOCK_LSN(be64_to_cpu(iclog->ic_header.h_lsn))); in xlog_sync()
1891 iclog->ic_bwritecnt = 2; in xlog_sync()
1900 dptr = (char *)&iclog->ic_header + count; in xlog_sync()
1910 iclog->ic_bwritecnt = 1; in xlog_sync()
1914 iclog->ic_header.h_crc = xlog_cksum(log, &iclog->ic_header, in xlog_sync()
1915 iclog->ic_datap, size); in xlog_sync()
1924 iclog->ic_header.h_crc &= cpu_to_le32(0xAAAAAAAA); in xlog_sync()
1925 iclog->ic_state |= XLOG_STATE_IOABORT; in xlog_sync()
1928 be64_to_cpu(iclog->ic_header.h_lsn)); in xlog_sync()
1932 bp->b_log_item = iclog; in xlog_sync()
1952 xlog_verify_iclog(log, iclog, count, true); in xlog_sync()
1967 bp = iclog->ic_log->l_xbuf; in xlog_sync()
1970 (char *)&iclog->ic_header + count, split); in xlog_sync()
1971 bp->b_log_item = iclog; in xlog_sync()
1996 xlog_in_core_t *iclog, *next_iclog; in xlog_dealloc_log() local
2005 iclog = log->l_iclog; in xlog_dealloc_log()
2007 xfs_buf_lock(iclog->ic_bp); in xlog_dealloc_log()
2008 xfs_buf_unlock(iclog->ic_bp); in xlog_dealloc_log()
2009 iclog = iclog->ic_next; in xlog_dealloc_log()
2022 iclog = log->l_iclog; in xlog_dealloc_log()
2024 xfs_buf_free(iclog->ic_bp); in xlog_dealloc_log()
2025 next_iclog = iclog->ic_next; in xlog_dealloc_log()
2026 kmem_free(iclog); in xlog_dealloc_log()
2027 iclog = next_iclog; in xlog_dealloc_log()
2042 struct xlog_in_core *iclog, in xlog_state_finish_copy() argument
2048 be32_add_cpu(&iclog->ic_header.h_num_logops, record_cnt); in xlog_state_finish_copy()
2049 iclog->ic_offset += copy_bytes; in xlog_state_finish_copy()
2316 struct xlog_in_core *iclog, in xlog_write_copy_finish() argument
2330 xlog_state_finish_copy(log, iclog, *record_cnt, *data_cnt); in xlog_write_copy_finish()
2333 return xlog_state_release_iclog(log, iclog); in xlog_write_copy_finish()
2339 if (iclog->ic_size - log_offset <= sizeof(xlog_op_header_t)) { in xlog_write_copy_finish()
2341 xlog_state_finish_copy(log, iclog, *record_cnt, *data_cnt); in xlog_write_copy_finish()
2346 xlog_state_want_sync(log, iclog); in xlog_write_copy_finish()
2350 return xlog_state_release_iclog(log, iclog); in xlog_write_copy_finish()
2352 *commit_iclog = iclog; in xlog_write_copy_finish()
2407 struct xlog_in_core *iclog = NULL; in xlog_write() local
2452 error = xlog_state_get_iclog_space(log, len, &iclog, ticket, in xlog_write()
2457 ASSERT(log_offset <= iclog->ic_size - 1); in xlog_write()
2458 ptr = iclog->ic_datap + log_offset; in xlog_write()
2462 *start_lsn = be64_to_cpu(iclog->ic_header.h_lsn); in xlog_write()
2502 iclog->ic_size-log_offset, in xlog_write()
2527 error = xlog_write_copy_finish(log, iclog, flags, in xlog_write()
2568 xlog_state_finish_copy(log, iclog, record_cnt, data_cnt); in xlog_write()
2570 return xlog_state_release_iclog(log, iclog); in xlog_write()
2573 *commit_iclog = iclog; in xlog_write()
2597 xlog_in_core_t *iclog; in xlog_state_clean_log() local
2600 iclog = log->l_iclog; in xlog_state_clean_log()
2602 if (iclog->ic_state == XLOG_STATE_DIRTY) { in xlog_state_clean_log()
2603 iclog->ic_state = XLOG_STATE_ACTIVE; in xlog_state_clean_log()
2604 iclog->ic_offset = 0; in xlog_state_clean_log()
2605 ASSERT(iclog->ic_callback == NULL); in xlog_state_clean_log()
2615 (be32_to_cpu(iclog->ic_header.h_num_logops) == in xlog_state_clean_log()
2626 iclog->ic_header.h_num_logops = 0; in xlog_state_clean_log()
2627 memset(iclog->ic_header.h_cycle_data, 0, in xlog_state_clean_log()
2628 sizeof(iclog->ic_header.h_cycle_data)); in xlog_state_clean_log()
2629 iclog->ic_header.h_lsn = 0; in xlog_state_clean_log()
2630 } else if (iclog->ic_state == XLOG_STATE_ACTIVE) in xlog_state_clean_log()
2634 iclog = iclog->ic_next; in xlog_state_clean_log()
2635 } while (iclog != log->l_iclog); in xlog_state_clean_log()
2702 xlog_in_core_t *iclog; in xlog_state_do_callback() local
2716 first_iclog = iclog = log->l_iclog; in xlog_state_do_callback()
2731 iclog = log->l_iclog; in xlog_state_do_callback()
2738 if (iclog->ic_state & in xlog_state_do_callback()
2740 iclog = iclog->ic_next; in xlog_state_do_callback()
2751 if (!(iclog->ic_state & XLOG_STATE_IOERROR)) { in xlog_state_do_callback()
2762 if (!(iclog->ic_state & in xlog_state_do_callback()
2788 be64_to_cpu(iclog->ic_header.h_lsn)) < 0) { in xlog_state_do_callback()
2789 iclog = iclog->ic_next; in xlog_state_do_callback()
2794 iclog->ic_state = XLOG_STATE_CALLBACK; in xlog_state_do_callback()
2815 be64_to_cpu(iclog->ic_header.h_lsn)) <= 0); in xlog_state_do_callback()
2816 if (iclog->ic_callback) in xlog_state_do_callback()
2818 be64_to_cpu(iclog->ic_header.h_lsn)); in xlog_state_do_callback()
2832 spin_lock(&iclog->ic_callback_lock); in xlog_state_do_callback()
2833 cb = iclog->ic_callback; in xlog_state_do_callback()
2835 iclog->ic_callback_tail = &(iclog->ic_callback); in xlog_state_do_callback()
2836 iclog->ic_callback = NULL; in xlog_state_do_callback()
2837 spin_unlock(&iclog->ic_callback_lock); in xlog_state_do_callback()
2844 spin_lock(&iclog->ic_callback_lock); in xlog_state_do_callback()
2845 cb = iclog->ic_callback; in xlog_state_do_callback()
2852 ASSERT(iclog->ic_callback == NULL); in xlog_state_do_callback()
2853 spin_unlock(&iclog->ic_callback_lock); in xlog_state_do_callback()
2854 if (!(iclog->ic_state & XLOG_STATE_IOERROR)) in xlog_state_do_callback()
2855 iclog->ic_state = XLOG_STATE_DIRTY; in xlog_state_do_callback()
2864 wake_up_all(&iclog->ic_force_wait); in xlog_state_do_callback()
2866 iclog = iclog->ic_next; in xlog_state_do_callback()
2867 } while (first_iclog != iclog); in xlog_state_do_callback()
2892 first_iclog = iclog = log->l_iclog; in xlog_state_do_callback()
2894 ASSERT(iclog->ic_state != XLOG_STATE_DO_CALLBACK); in xlog_state_do_callback()
2904 if (iclog->ic_state == XLOG_STATE_WANT_SYNC || in xlog_state_do_callback()
2905 iclog->ic_state & XLOG_STATE_SYNCING || in xlog_state_do_callback()
2906 iclog->ic_state == XLOG_STATE_DONE_SYNC || in xlog_state_do_callback()
2907 iclog->ic_state == XLOG_STATE_IOERROR ) in xlog_state_do_callback()
2909 iclog = iclog->ic_next; in xlog_state_do_callback()
2910 } while (first_iclog != iclog); in xlog_state_do_callback()
2938 xlog_in_core_t *iclog, in xlog_state_done_syncing() argument
2941 struct xlog *log = iclog->ic_log; in xlog_state_done_syncing()
2945 ASSERT(iclog->ic_state == XLOG_STATE_SYNCING || in xlog_state_done_syncing()
2946 iclog->ic_state == XLOG_STATE_IOERROR); in xlog_state_done_syncing()
2947 ASSERT(atomic_read(&iclog->ic_refcnt) == 0); in xlog_state_done_syncing()
2948 ASSERT(iclog->ic_bwritecnt == 1 || iclog->ic_bwritecnt == 2); in xlog_state_done_syncing()
2957 if (iclog->ic_state != XLOG_STATE_IOERROR) { in xlog_state_done_syncing()
2958 if (--iclog->ic_bwritecnt == 1) { in xlog_state_done_syncing()
2962 iclog->ic_state = XLOG_STATE_DONE_SYNC; in xlog_state_done_syncing()
2970 wake_up_all(&iclog->ic_write_wait); in xlog_state_done_syncing()
2972 xlog_state_do_callback(log, aborted, iclog); /* also cleans log */ in xlog_state_done_syncing()
3005 xlog_in_core_t *iclog; in xlog_state_get_iclog_space() local
3015 iclog = log->l_iclog; in xlog_state_get_iclog_space()
3016 if (iclog->ic_state != XLOG_STATE_ACTIVE) { in xlog_state_get_iclog_space()
3024 head = &iclog->ic_header; in xlog_state_get_iclog_space()
3026 atomic_inc(&iclog->ic_refcnt); /* prevents sync */ in xlog_state_get_iclog_space()
3027 log_offset = iclog->ic_offset; in xlog_state_get_iclog_space()
3054 if (iclog->ic_size - iclog->ic_offset < 2*sizeof(xlog_op_header_t)) { in xlog_state_get_iclog_space()
3055 xlog_state_switch_iclogs(log, iclog, iclog->ic_size); in xlog_state_get_iclog_space()
3064 if (!atomic_add_unless(&iclog->ic_refcnt, -1, 1)) { in xlog_state_get_iclog_space()
3067 error = xlog_state_release_iclog(log, iclog); in xlog_state_get_iclog_space()
3082 if (len <= iclog->ic_size - iclog->ic_offset) { in xlog_state_get_iclog_space()
3084 iclog->ic_offset += len; in xlog_state_get_iclog_space()
3087 xlog_state_switch_iclogs(log, iclog, iclog->ic_size); in xlog_state_get_iclog_space()
3089 *iclogp = iclog; in xlog_state_get_iclog_space()
3091 ASSERT(iclog->ic_offset <= iclog->ic_size); in xlog_state_get_iclog_space()
3195 struct xlog_in_core *iclog) in xlog_state_release_iclog() argument
3199 if (iclog->ic_state & XLOG_STATE_IOERROR) in xlog_state_release_iclog()
3202 ASSERT(atomic_read(&iclog->ic_refcnt) > 0); in xlog_state_release_iclog()
3203 if (!atomic_dec_and_lock(&iclog->ic_refcnt, &log->l_icloglock)) in xlog_state_release_iclog()
3206 if (iclog->ic_state & XLOG_STATE_IOERROR) { in xlog_state_release_iclog()
3210 ASSERT(iclog->ic_state == XLOG_STATE_ACTIVE || in xlog_state_release_iclog()
3211 iclog->ic_state == XLOG_STATE_WANT_SYNC); in xlog_state_release_iclog()
3213 if (iclog->ic_state == XLOG_STATE_WANT_SYNC) { in xlog_state_release_iclog()
3217 iclog->ic_state = XLOG_STATE_SYNCING; in xlog_state_release_iclog()
3218 iclog->ic_header.h_tail_lsn = cpu_to_be64(tail_lsn); in xlog_state_release_iclog()
3219 xlog_verify_tail_lsn(log, iclog, tail_lsn); in xlog_state_release_iclog()
3232 return xlog_sync(log, iclog); in xlog_state_release_iclog()
3247 struct xlog_in_core *iclog, in xlog_state_switch_iclogs() argument
3250 ASSERT(iclog->ic_state == XLOG_STATE_ACTIVE); in xlog_state_switch_iclogs()
3252 eventual_size = iclog->ic_offset; in xlog_state_switch_iclogs()
3253 iclog->ic_state = XLOG_STATE_WANT_SYNC; in xlog_state_switch_iclogs()
3254 iclog->ic_header.h_prev_block = cpu_to_be32(log->l_prev_block); in xlog_state_switch_iclogs()
3283 ASSERT(iclog == log->l_iclog); in xlog_state_switch_iclogs()
3284 log->l_iclog = iclog->ic_next; in xlog_state_switch_iclogs()
3320 struct xlog_in_core *iclog; in xfs_log_force() local
3329 iclog = log->l_iclog; in xfs_log_force()
3330 if (iclog->ic_state & XLOG_STATE_IOERROR) in xfs_log_force()
3333 if (iclog->ic_state == XLOG_STATE_DIRTY || in xfs_log_force()
3334 (iclog->ic_state == XLOG_STATE_ACTIVE && in xfs_log_force()
3335 atomic_read(&iclog->ic_refcnt) == 0 && iclog->ic_offset == 0)) { in xfs_log_force()
3344 iclog = iclog->ic_prev; in xfs_log_force()
3345 if (iclog->ic_state == XLOG_STATE_ACTIVE || in xfs_log_force()
3346 iclog->ic_state == XLOG_STATE_DIRTY) in xfs_log_force()
3348 } else if (iclog->ic_state == XLOG_STATE_ACTIVE) { in xfs_log_force()
3349 if (atomic_read(&iclog->ic_refcnt) == 0) { in xfs_log_force()
3357 atomic_inc(&iclog->ic_refcnt); in xfs_log_force()
3358 lsn = be64_to_cpu(iclog->ic_header.h_lsn); in xfs_log_force()
3359 xlog_state_switch_iclogs(log, iclog, 0); in xfs_log_force()
3362 if (xlog_state_release_iclog(log, iclog)) in xfs_log_force()
3366 if (be64_to_cpu(iclog->ic_header.h_lsn) != lsn || in xfs_log_force()
3367 iclog->ic_state == XLOG_STATE_DIRTY) in xfs_log_force()
3377 xlog_state_switch_iclogs(log, iclog, 0); in xfs_log_force()
3390 if (iclog->ic_state & XLOG_STATE_IOERROR) in xfs_log_force()
3393 xlog_wait(&iclog->ic_force_wait, &log->l_icloglock); in xfs_log_force()
3394 if (iclog->ic_state & XLOG_STATE_IOERROR) in xfs_log_force()
3415 struct xlog_in_core *iclog; in __xfs_log_force_lsn() local
3418 iclog = log->l_iclog; in __xfs_log_force_lsn()
3419 if (iclog->ic_state & XLOG_STATE_IOERROR) in __xfs_log_force_lsn()
3422 while (be64_to_cpu(iclog->ic_header.h_lsn) != lsn) { in __xfs_log_force_lsn()
3423 iclog = iclog->ic_next; in __xfs_log_force_lsn()
3424 if (iclog == log->l_iclog) in __xfs_log_force_lsn()
3428 if (iclog->ic_state == XLOG_STATE_DIRTY) in __xfs_log_force_lsn()
3431 if (iclog->ic_state == XLOG_STATE_ACTIVE) { in __xfs_log_force_lsn()
3448 (iclog->ic_prev->ic_state & in __xfs_log_force_lsn()
3450 ASSERT(!(iclog->ic_state & XLOG_STATE_IOERROR)); in __xfs_log_force_lsn()
3454 xlog_wait(&iclog->ic_prev->ic_write_wait, in __xfs_log_force_lsn()
3458 atomic_inc(&iclog->ic_refcnt); in __xfs_log_force_lsn()
3459 xlog_state_switch_iclogs(log, iclog, 0); in __xfs_log_force_lsn()
3461 if (xlog_state_release_iclog(log, iclog)) in __xfs_log_force_lsn()
3469 (iclog->ic_state & (XLOG_STATE_ACTIVE | XLOG_STATE_DIRTY))) in __xfs_log_force_lsn()
3472 if (iclog->ic_state & XLOG_STATE_IOERROR) in __xfs_log_force_lsn()
3476 xlog_wait(&iclog->ic_force_wait, &log->l_icloglock); in __xfs_log_force_lsn()
3477 if (iclog->ic_state & XLOG_STATE_IOERROR) in __xfs_log_force_lsn()
3533 struct xlog_in_core *iclog) in xlog_state_want_sync() argument
3537 if (iclog->ic_state == XLOG_STATE_ACTIVE) { in xlog_state_want_sync()
3538 xlog_state_switch_iclogs(log, iclog, 0); in xlog_state_want_sync()
3540 ASSERT(iclog->ic_state & in xlog_state_want_sync()
3782 struct xlog_in_core *iclog, in xlog_verify_tail_lsn() argument
3790 if (blocks < BTOBB(iclog->ic_offset)+BTOBB(log->l_iclog_hsize)) in xlog_verify_tail_lsn()
3799 if (blocks < BTOBB(iclog->ic_offset) + 1) in xlog_verify_tail_lsn()
3822 struct xlog_in_core *iclog, in xlog_verify_iclog() argument
3846 if (iclog->ic_header.h_magicno != cpu_to_be32(XLOG_HEADER_MAGIC_NUM)) in xlog_verify_iclog()
3849 base_ptr = ptr = &iclog->ic_header; in xlog_verify_iclog()
3850 p = &iclog->ic_header; in xlog_verify_iclog()
3858 len = be32_to_cpu(iclog->ic_header.h_num_logops); in xlog_verify_iclog()
3859 base_ptr = ptr = iclog->ic_datap; in xlog_verify_iclog()
3861 xhdr = iclog->ic_data; in xlog_verify_iclog()
3871 idx = BTOBBT((char *)&ophead->oh_clientid - iclog->ic_datap); in xlog_verify_iclog()
3879 iclog->ic_header.h_cycle_data[idx]); in xlog_verify_iclog()
3895 (uintptr_t)iclog->ic_datap); in xlog_verify_iclog()
3901 op_len = be32_to_cpu(iclog->ic_header.h_cycle_data[idx]); in xlog_verify_iclog()
3916 xlog_in_core_t *iclog, *ic; in xlog_state_ioerror() local
3918 iclog = log->l_iclog; in xlog_state_ioerror()
3919 if (! (iclog->ic_state & XLOG_STATE_IOERROR)) { in xlog_state_ioerror()
3924 ic = iclog; in xlog_state_ioerror()
3928 } while (ic != iclog); in xlog_state_ioerror()
4032 xlog_in_core_t *iclog; in xfs_log_force_umount() local
4035 iclog = log->l_iclog; in xfs_log_force_umount()
4037 ASSERT(iclog->ic_callback == 0); in xfs_log_force_umount()
4038 iclog = iclog->ic_next; in xfs_log_force_umount()
4039 } while (iclog != log->l_iclog); in xfs_log_force_umount()
4051 xlog_in_core_t *iclog; in xlog_iclogs_empty() local
4053 iclog = log->l_iclog; in xlog_iclogs_empty()
4058 if (iclog->ic_header.h_num_logops) in xlog_iclogs_empty()
4060 iclog = iclog->ic_next; in xlog_iclogs_empty()
4061 } while (iclog != log->l_iclog); in xlog_iclogs_empty()