Lines Matching full:work
92 struct cm_work *work);
184 struct delayed_work work; member
195 struct cm_work work; member
267 static void cm_work_handler(struct work_struct *work);
693 __be32 remote_id = timewait_info->work.remote_id; in cm_insert_remote_id()
699 if (be32_lt(remote_id, cur_timewait_info->work.remote_id)) in cm_insert_remote_id()
701 else if (be32_gt(remote_id, cur_timewait_info->work.remote_id)) in cm_insert_remote_id()
727 if (be32_lt(remote_id, timewait_info->work.remote_id)) in cm_find_remote_id()
729 else if (be32_gt(remote_id, timewait_info->work.remote_id)) in cm_find_remote_id()
736 res = cm_acquire_id(timewait_info->work.local_id, in cm_find_remote_id()
737 timewait_info->work.remote_id); in cm_find_remote_id()
871 struct cm_work *work; in cm_dequeue_work() local
876 work = list_entry(cm_id_priv->work_list.next, struct cm_work, list); in cm_dequeue_work()
877 list_del(&work->list); in cm_dequeue_work()
878 return work; in cm_dequeue_work()
881 static void cm_free_work(struct cm_work *work) in cm_free_work() argument
883 if (work->mad_recv_wc) in cm_free_work()
884 ib_free_recv_mad(work->mad_recv_wc); in cm_free_work()
885 kfree(work); in cm_free_work()
889 struct cm_work *work) in cm_queue_work_unlock() argument
903 list_add_tail(&work->list, &cm_id_priv->work_list); in cm_queue_work_unlock()
915 cm_process_work(cm_id_priv, work); in cm_queue_work_unlock()
966 timewait_info->work.local_id = local_id; in cm_create_timewait_info()
967 INIT_DELAYED_WORK(&timewait_info->work.work, cm_work_handler); in cm_create_timewait_info()
968 timewait_info->work.cm_event.event = IB_CM_TIMEWAIT_EXIT; in cm_create_timewait_info()
1000 queue_delayed_work(cm.wq, &cm_id_priv->timewait_info->work.work, in cm_enter_timewait()
1005 * The timewait_info is converted into a work and gets freed during in cm_enter_timewait()
1008 BUILD_BUG_ON(offsetof(struct cm_timewait_info, work) != 0); in cm_enter_timewait()
1031 struct cm_work *work; in cm_destroy_id() local
1139 while ((work = cm_dequeue_work(cm_id_priv)) != NULL) in cm_destroy_id()
1140 cm_free_work(work); in cm_destroy_id()
1377 /* Work-around until there's a way to obtain remote LID info */ in cm_format_req()
1735 static u16 cm_get_bth_pkey(struct cm_work *work) in cm_get_bth_pkey() argument
1737 struct ib_device *ib_dev = work->port->cm_dev->ib_device; in cm_get_bth_pkey()
1738 u32 port_num = work->port->port_num; in cm_get_bth_pkey()
1739 u16 pkey_index = work->mad_recv_wc->wc->pkey_index; in cm_get_bth_pkey()
1759 * @work: Work completion
1762 static void cm_opa_to_ib_sgid(struct cm_work *work, in cm_opa_to_ib_sgid() argument
1765 struct ib_device *dev = work->port->cm_dev->ib_device; in cm_opa_to_ib_sgid()
1766 u32 port_num = work->port->port_num; in cm_opa_to_ib_sgid()
1782 static void cm_format_req_event(struct cm_work *work, in cm_format_req_event() argument
1789 req_msg = (struct cm_req_msg *)work->mad_recv_wc->recv_buf.mad; in cm_format_req_event()
1790 param = &work->cm_event.param.req_rcvd; in cm_format_req_event()
1792 param->bth_pkey = cm_get_bth_pkey(work); in cm_format_req_event()
1794 param->primary_path = &work->path[0]; in cm_format_req_event()
1795 cm_opa_to_ib_sgid(work, param->primary_path); in cm_format_req_event()
1797 param->alternate_path = &work->path[1]; in cm_format_req_event()
1798 cm_opa_to_ib_sgid(work, param->alternate_path); in cm_format_req_event()
1822 work->cm_event.private_data = in cm_format_req_event()
1827 struct cm_work *work) in cm_process_work() argument
1832 ret = cm_id_priv->id.cm_handler(&cm_id_priv->id, &work->cm_event); in cm_process_work()
1833 cm_free_work(work); in cm_process_work()
1837 work = cm_dequeue_work(cm_id_priv); in cm_process_work()
1839 if (!work) in cm_process_work()
1843 &work->cm_event); in cm_process_work()
1844 cm_free_work(work); in cm_process_work()
1916 static void cm_dup_req_handler(struct cm_work *work, in cm_dup_req_handler() argument
1923 &work->port->counters[CM_RECV_DUPLICATES][CM_REQ_COUNTER]); in cm_dup_req_handler()
1933 ret = cm_alloc_response_msg(work->port, work->mad_recv_wc, &msg); in cm_dup_req_handler()
1965 static struct cm_id_private *cm_match_req(struct cm_work *work, in cm_match_req() argument
1972 req_msg = (struct cm_req_msg *)work->mad_recv_wc->recv_buf.mad; in cm_match_req()
1978 cur_cm_id_priv = cm_acquire_id(timewait_info->work.local_id, in cm_match_req()
1979 timewait_info->work.remote_id); in cm_match_req()
1982 cm_dup_req_handler(work, cur_cm_id_priv); in cm_match_req()
1992 cur_cm_id_priv = cm_acquire_id(timewait_info->work.local_id, in cm_match_req()
1993 timewait_info->work.remote_id); in cm_match_req()
1996 cm_issue_rej(work->port, work->mad_recv_wc, in cm_match_req()
2013 cm_issue_rej(work->port, work->mad_recv_wc, in cm_match_req()
2023 * Work-around for inter-subnet connections. If the LIDs are permissive,
2025 * in the work completion.
2058 static int cm_req_handler(struct cm_work *work) in cm_req_handler() argument
2066 req_msg = (struct cm_req_msg *)work->mad_recv_wc->recv_buf.mad; in cm_req_handler()
2069 cm_alloc_id_priv(work->port->cm_dev->ib_device, NULL, NULL); in cm_req_handler()
2094 ret = cm_init_av_for_response(work->port, work->mad_recv_wc->wc, in cm_req_handler()
2095 work->mad_recv_wc->recv_buf.grh, in cm_req_handler()
2106 cm_id_priv->timewait_info->work.remote_id = cm_id_priv->id.remote_id; in cm_req_handler()
2117 listen_cm_id_priv = cm_match_req(work, cm_id_priv); in cm_req_handler()
2125 memset(&work->path[0], 0, sizeof(work->path[0])); in cm_req_handler()
2127 memset(&work->path[1], 0, sizeof(work->path[1])); in cm_req_handler()
2132 work->path[0].rec_type = in cm_req_handler()
2135 cm_process_routed_req(req_msg, work->mad_recv_wc->wc); in cm_req_handler()
2137 work->port->cm_dev->ib_device, work->port->port_num, in cm_req_handler()
2138 &work->path[0], in cm_req_handler()
2143 work->path[1].rec_type = work->path[0].rec_type; in cm_req_handler()
2144 cm_format_paths_from_req(req_msg, &work->path[0], in cm_req_handler()
2145 &work->path[1], work->mad_recv_wc->wc); in cm_req_handler()
2147 sa_path_set_dmac(&work->path[0], in cm_req_handler()
2149 work->path[0].hop_limit = grh->hop_limit; in cm_req_handler()
2153 ret = cm_init_av_by_path(&work->path[0], gid_attr, &cm_id_priv->av); in cm_req_handler()
2157 err = rdma_query_gid(work->port->cm_dev->ib_device, in cm_req_handler()
2158 work->port->port_num, 0, in cm_req_handler()
2159 &work->path[0].sgid); in cm_req_handler()
2165 &work->path[0].sgid, in cm_req_handler()
2166 sizeof(work->path[0].sgid), in cm_req_handler()
2175 ret = cm_init_av_by_path(&work->path[1], NULL, in cm_req_handler()
2180 &work->path[0].sgid, in cm_req_handler()
2181 sizeof(work->path[0].sgid), NULL, 0); in cm_req_handler()
2188 cm_format_req_event(work, cm_id_priv, &listen_cm_id_priv->id); in cm_req_handler()
2196 cm_queue_work_unlock(cm_id_priv, work); in cm_req_handler()
2382 static void cm_format_rep_event(struct cm_work *work, enum ib_qp_type qp_type) in cm_format_rep_event() argument
2387 rep_msg = (struct cm_rep_msg *)work->mad_recv_wc->recv_buf.mad; in cm_format_rep_event()
2388 param = &work->cm_event.param.rep_rcvd; in cm_format_rep_event()
2406 work->cm_event.private_data = in cm_format_rep_event()
2410 static void cm_dup_rep_handler(struct cm_work *work) in cm_dup_rep_handler() argument
2417 rep_msg = (struct cm_rep_msg *) work->mad_recv_wc->recv_buf.mad; in cm_dup_rep_handler()
2425 &work->port->counters[CM_RECV_DUPLICATES][CM_REP_COUNTER]); in cm_dup_rep_handler()
2426 ret = cm_alloc_response_msg(work->port, work->mad_recv_wc, &msg); in cm_dup_rep_handler()
2455 static int cm_rep_handler(struct cm_work *work) in cm_rep_handler() argument
2463 rep_msg = (struct cm_rep_msg *)work->mad_recv_wc->recv_buf.mad; in cm_rep_handler()
2467 cm_dup_rep_handler(work); in cm_rep_handler()
2473 cm_format_rep_event(work, cm_id_priv->qp_type); in cm_rep_handler()
2490 cm_id_priv->timewait_info->work.remote_id = in cm_rep_handler()
2510 cur_cm_id_priv = cm_acquire_id(timewait_info->work.local_id, in cm_rep_handler()
2511 timewait_info->work.remote_id); in cm_rep_handler()
2515 cm_issue_rej(work->port, work->mad_recv_wc, in cm_rep_handler()
2552 cm_queue_work_unlock(cm_id_priv, work); in cm_rep_handler()
2560 static int cm_establish_handler(struct cm_work *work) in cm_establish_handler() argument
2565 cm_id_priv = cm_acquire_id(work->local_id, work->remote_id); in cm_establish_handler()
2576 cm_queue_work_unlock(cm_id_priv, work); in cm_establish_handler()
2583 static int cm_rtu_handler(struct cm_work *work) in cm_rtu_handler() argument
2588 rtu_msg = (struct cm_rtu_msg *)work->mad_recv_wc->recv_buf.mad; in cm_rtu_handler()
2595 work->cm_event.private_data = in cm_rtu_handler()
2602 atomic_long_inc(&work->port->counters[CM_RECV_DUPLICATES] in cm_rtu_handler()
2609 cm_queue_work_unlock(cm_id_priv, work); in cm_rtu_handler()
2796 static int cm_dreq_handler(struct cm_work *work) in cm_dreq_handler() argument
2802 dreq_msg = (struct cm_dreq_msg *)work->mad_recv_wc->recv_buf.mad; in cm_dreq_handler()
2807 atomic_long_inc(&work->port->counters[CM_RECV_DUPLICATES] in cm_dreq_handler()
2809 cm_issue_drep(work->port, work->mad_recv_wc); in cm_dreq_handler()
2816 work->cm_event.private_data = in cm_dreq_handler()
2836 atomic_long_inc(&work->port->counters[CM_RECV_DUPLICATES] in cm_dreq_handler()
2838 msg = cm_alloc_response_msg_no_ah(work->port, work->mad_recv_wc); in cm_dreq_handler()
2847 if (cm_create_response_msg_ah(work->port, work->mad_recv_wc, msg) || in cm_dreq_handler()
2852 atomic_long_inc(&work->port->counters[CM_RECV_DUPLICATES] in cm_dreq_handler()
2861 cm_queue_work_unlock(cm_id_priv, work); in cm_dreq_handler()
2869 static int cm_drep_handler(struct cm_work *work) in cm_drep_handler() argument
2874 drep_msg = (struct cm_drep_msg *)work->mad_recv_wc->recv_buf.mad; in cm_drep_handler()
2881 work->cm_event.private_data = in cm_drep_handler()
2893 cm_queue_work_unlock(cm_id_priv, work); in cm_drep_handler()
2972 static void cm_format_rej_event(struct cm_work *work) in cm_format_rej_event() argument
2977 rej_msg = (struct cm_rej_msg *)work->mad_recv_wc->recv_buf.mad; in cm_format_rej_event()
2978 param = &work->cm_event.param.rej_rcvd; in cm_format_rej_event()
2982 work->cm_event.private_data = in cm_format_rej_event()
3010 static int cm_rej_handler(struct cm_work *work) in cm_rej_handler() argument
3015 rej_msg = (struct cm_rej_msg *)work->mad_recv_wc->recv_buf.mad; in cm_rej_handler()
3020 cm_format_rej_event(work); in cm_rej_handler()
3059 cm_queue_work_unlock(cm_id_priv, work); in cm_rej_handler()
3164 static int cm_mra_handler(struct cm_work *work) in cm_mra_handler() argument
3170 mra_msg = (struct cm_mra_msg *)work->mad_recv_wc->recv_buf.mad; in cm_mra_handler()
3175 work->cm_event.private_data = in cm_mra_handler()
3177 work->cm_event.param.mra_rcvd.service_timeout = in cm_mra_handler()
3205 &work->port->counters[CM_RECV_DUPLICATES] in cm_mra_handler()
3213 atomic_long_inc(&work->port->counters[CM_RECV_DUPLICATES] in cm_mra_handler()
3223 cm_queue_work_unlock(cm_id_priv, work); in cm_mra_handler()
3277 static int cm_lap_handler(struct cm_work *work) in cm_lap_handler() argument
3290 if (rdma_protocol_roce(work->port->cm_dev->ib_device, in cm_lap_handler()
3291 work->port->port_num)) in cm_lap_handler()
3295 lap_msg = (struct cm_lap_msg *)work->mad_recv_wc->recv_buf.mad; in cm_lap_handler()
3302 param = &work->cm_event.param.lap_rcvd; in cm_lap_handler()
3303 memset(&work->path[0], 0, sizeof(work->path[1])); in cm_lap_handler()
3304 cm_path_set_rec_type(work->port->cm_dev->ib_device, in cm_lap_handler()
3305 work->port->port_num, &work->path[0], in cm_lap_handler()
3308 param->alternate_path = &work->path[0]; in cm_lap_handler()
3310 work->cm_event.private_data = in cm_lap_handler()
3313 ret = ib_init_ah_attr_from_wc(work->port->cm_dev->ib_device, in cm_lap_handler()
3314 work->port->port_num, in cm_lap_handler()
3315 work->mad_recv_wc->wc, in cm_lap_handler()
3316 work->mad_recv_wc->recv_buf.grh, in cm_lap_handler()
3328 cm_init_av_for_lap(work->port, work->mad_recv_wc->wc, in cm_lap_handler()
3340 atomic_long_inc(&work->port->counters[CM_RECV_DUPLICATES] in cm_lap_handler()
3342 msg = cm_alloc_response_msg_no_ah(work->port, work->mad_recv_wc); in cm_lap_handler()
3353 if (cm_create_response_msg_ah(work->port, work->mad_recv_wc, msg) || in cm_lap_handler()
3358 atomic_long_inc(&work->port->counters[CM_RECV_DUPLICATES] in cm_lap_handler()
3367 cm_queue_work_unlock(cm_id_priv, work); in cm_lap_handler()
3375 static int cm_apr_handler(struct cm_work *work) in cm_apr_handler() argument
3383 if (rdma_protocol_roce(work->port->cm_dev->ib_device, in cm_apr_handler()
3384 work->port->port_num)) in cm_apr_handler()
3387 apr_msg = (struct cm_apr_msg *)work->mad_recv_wc->recv_buf.mad; in cm_apr_handler()
3394 work->cm_event.param.apr_rcvd.ap_status = in cm_apr_handler()
3396 work->cm_event.param.apr_rcvd.apr_info = in cm_apr_handler()
3398 work->cm_event.param.apr_rcvd.info_len = in cm_apr_handler()
3400 work->cm_event.private_data = in cm_apr_handler()
3412 cm_queue_work_unlock(cm_id_priv, work); in cm_apr_handler()
3419 static int cm_timewait_handler(struct cm_work *work) in cm_timewait_handler() argument
3424 timewait_info = container_of(work, struct cm_timewait_info, work); in cm_timewait_handler()
3429 cm_id_priv = cm_acquire_id(timewait_info->work.local_id, in cm_timewait_handler()
3430 timewait_info->work.remote_id); in cm_timewait_handler()
3441 cm_queue_work_unlock(cm_id_priv, work); in cm_timewait_handler()
3520 static void cm_format_sidr_req_event(struct cm_work *work, in cm_format_sidr_req_event() argument
3528 work->mad_recv_wc->recv_buf.mad; in cm_format_sidr_req_event()
3529 param = &work->cm_event.param.sidr_req_rcvd; in cm_format_sidr_req_event()
3534 param->bth_pkey = cm_get_bth_pkey(work); in cm_format_sidr_req_event()
3535 param->port = work->port->port_num; in cm_format_sidr_req_event()
3537 work->cm_event.private_data = in cm_format_sidr_req_event()
3541 static int cm_sidr_req_handler(struct cm_work *work) in cm_sidr_req_handler() argument
3549 cm_alloc_id_priv(work->port->cm_dev->ib_device, NULL, NULL); in cm_sidr_req_handler()
3555 work->mad_recv_wc->recv_buf.mad; in cm_sidr_req_handler()
3563 wc = work->mad_recv_wc->wc; in cm_sidr_req_handler()
3565 ret = cm_init_av_for_response(work->port, work->mad_recv_wc->wc, in cm_sidr_req_handler()
3566 work->mad_recv_wc->recv_buf.grh, in cm_sidr_req_handler()
3575 atomic_long_inc(&work->port->counters[CM_RECV_DUPLICATES] in cm_sidr_req_handler()
3600 cm_format_sidr_req_event(work, cm_id_priv, &listen_cm_id_priv->id); in cm_sidr_req_handler()
3601 ret = cm_id_priv->id.cm_handler(&cm_id_priv->id, &work->cm_event); in cm_sidr_req_handler()
3602 cm_free_work(work); in cm_sidr_req_handler()
3697 static void cm_format_sidr_rep_event(struct cm_work *work, in cm_format_sidr_rep_event() argument
3704 work->mad_recv_wc->recv_buf.mad; in cm_format_sidr_rep_event()
3705 param = &work->cm_event.param.sidr_rep_rcvd; in cm_format_sidr_rep_event()
3714 work->cm_event.private_data = in cm_format_sidr_rep_event()
3718 static int cm_sidr_rep_handler(struct cm_work *work) in cm_sidr_rep_handler() argument
3724 work->mad_recv_wc->recv_buf.mad; in cm_sidr_rep_handler()
3739 cm_format_sidr_rep_event(work, cm_id_priv); in cm_sidr_rep_handler()
3740 cm_process_work(cm_id_priv, work); in cm_sidr_rep_handler()
3839 struct cm_work *work = container_of(_work, struct cm_work, work.work); in cm_work_handler() local
3842 switch (work->cm_event.event) { in cm_work_handler()
3844 ret = cm_req_handler(work); in cm_work_handler()
3847 ret = cm_mra_handler(work); in cm_work_handler()
3850 ret = cm_rej_handler(work); in cm_work_handler()
3853 ret = cm_rep_handler(work); in cm_work_handler()
3856 ret = cm_rtu_handler(work); in cm_work_handler()
3859 ret = cm_establish_handler(work); in cm_work_handler()
3862 ret = cm_dreq_handler(work); in cm_work_handler()
3865 ret = cm_drep_handler(work); in cm_work_handler()
3868 ret = cm_sidr_req_handler(work); in cm_work_handler()
3871 ret = cm_sidr_rep_handler(work); in cm_work_handler()
3874 ret = cm_lap_handler(work); in cm_work_handler()
3877 ret = cm_apr_handler(work); in cm_work_handler()
3880 ret = cm_timewait_handler(work); in cm_work_handler()
3883 trace_icm_handler_err(work->cm_event.event); in cm_work_handler()
3888 cm_free_work(work); in cm_work_handler()
3894 struct cm_work *work; in cm_establish() local
3903 work = kmalloc(sizeof *work, GFP_ATOMIC); in cm_establish()
3904 if (!work) in cm_establish()
3925 kfree(work); in cm_establish()
3931 * can execute this work item. To prevent potential deadlock, in cm_establish()
3935 INIT_DELAYED_WORK(&work->work, cm_work_handler); in cm_establish()
3936 work->local_id = cm_id->local_id; in cm_establish()
3937 work->remote_id = cm_id->remote_id; in cm_establish()
3938 work->mad_recv_wc = NULL; in cm_establish()
3939 work->cm_event.event = IB_CM_USER_ESTABLISHED; in cm_establish()
3944 queue_delayed_work(cm.wq, &work->work, 0); in cm_establish()
3946 kfree(work); in cm_establish()
3998 struct cm_work *work; in cm_recv_handler() local
4051 work = kmalloc(struct_size(work, path, paths), GFP_KERNEL); in cm_recv_handler()
4052 if (!work) { in cm_recv_handler()
4057 INIT_DELAYED_WORK(&work->work, cm_work_handler); in cm_recv_handler()
4058 work->cm_event.event = event; in cm_recv_handler()
4059 work->mad_recv_wc = mad_recv_wc; in cm_recv_handler()
4060 work->port = port; in cm_recv_handler()
4065 queue_delayed_work(cm.wq, &work->work, 0); in cm_recv_handler()
4071 kfree(work); in cm_recv_handler()
4459 * The above ensures no call paths from the work are running, in cm_remove_one()
4512 cancel_delayed_work(&timewait_info->work.work); in ib_cm_cleanup()