Lines Matching refs:mci
64 struct mem_ctl_info *mci = dimm->mci; in edac_dimm_info_location() local
68 for (i = 0; i < mci->n_layers; i++) { in edac_dimm_info_location()
70 edac_layer_name[mci->layers[i].type], in edac_dimm_info_location()
100 dimm->mci->csbased ? "rank" : "dimm", in edac_mc_dump_dimm()
117 edac_dbg(4, " csrow->mci = %p\n", csrow->mci); in edac_mc_dump_csrow()
120 static void edac_mc_dump_mci(struct mem_ctl_info *mci) in edac_mc_dump_mci() argument
122 edac_dbg(3, "\tmci = %p\n", mci); in edac_mc_dump_mci()
123 edac_dbg(3, "\tmci->mtype_cap = %lx\n", mci->mtype_cap); in edac_mc_dump_mci()
124 edac_dbg(3, "\tmci->edac_ctl_cap = %lx\n", mci->edac_ctl_cap); in edac_mc_dump_mci()
125 edac_dbg(3, "\tmci->edac_cap = %lx\n", mci->edac_cap); in edac_mc_dump_mci()
126 edac_dbg(4, "\tmci->edac_check = %p\n", mci->edac_check); in edac_mc_dump_mci()
128 mci->nr_csrows, mci->csrows); in edac_mc_dump_mci()
130 mci->tot_dimms, mci->dimms); in edac_mc_dump_mci()
131 edac_dbg(3, "\tdev = %p\n", mci->pdev); in edac_mc_dump_mci()
133 mci->mod_name, mci->ctl_name); in edac_mc_dump_mci()
134 edac_dbg(3, "\tpvt_info = %p\n\n", mci->pvt_info); in edac_mc_dump_mci()
172 static void _edac_mc_free(struct mem_ctl_info *mci) in _edac_mc_free() argument
174 put_device(&mci->dev); in _edac_mc_free()
179 struct mem_ctl_info *mci = container_of(dev, struct mem_ctl_info, dev); in mci_release() local
183 if (mci->dimms) { in mci_release()
184 for (i = 0; i < mci->tot_dimms; i++) in mci_release()
185 kfree(mci->dimms[i]); in mci_release()
186 kfree(mci->dimms); in mci_release()
189 if (mci->csrows) { in mci_release()
190 for (row = 0; row < mci->nr_csrows; row++) { in mci_release()
191 csr = mci->csrows[row]; in mci_release()
196 for (chn = 0; chn < mci->num_cschannel; chn++) in mci_release()
202 kfree(mci->csrows); in mci_release()
204 kfree(mci->pvt_info); in mci_release()
205 kfree(mci->layers); in mci_release()
206 kfree(mci); in mci_release()
209 static int edac_mc_alloc_csrows(struct mem_ctl_info *mci) in edac_mc_alloc_csrows() argument
211 unsigned int tot_channels = mci->num_cschannel; in edac_mc_alloc_csrows()
212 unsigned int tot_csrows = mci->nr_csrows; in edac_mc_alloc_csrows()
218 mci->csrows = kcalloc(tot_csrows, sizeof(*mci->csrows), GFP_KERNEL); in edac_mc_alloc_csrows()
219 if (!mci->csrows) in edac_mc_alloc_csrows()
225 csr = kzalloc(sizeof(**mci->csrows), GFP_KERNEL); in edac_mc_alloc_csrows()
229 mci->csrows[row] = csr; in edac_mc_alloc_csrows()
231 csr->mci = mci; in edac_mc_alloc_csrows()
254 static int edac_mc_alloc_dimms(struct mem_ctl_info *mci) in edac_mc_alloc_dimms() argument
264 mci->dimms = kcalloc(mci->tot_dimms, sizeof(*mci->dimms), GFP_KERNEL); in edac_mc_alloc_dimms()
265 if (!mci->dimms) in edac_mc_alloc_dimms()
271 for (idx = 0; idx < mci->tot_dimms; idx++) { in edac_mc_alloc_dimms()
276 chan = mci->csrows[row]->channels[chn]; in edac_mc_alloc_dimms()
278 dimm = kzalloc(sizeof(**mci->dimms), GFP_KERNEL); in edac_mc_alloc_dimms()
281 mci->dimms[idx] = dimm; in edac_mc_alloc_dimms()
282 dimm->mci = mci; in edac_mc_alloc_dimms()
290 n = scnprintf(p, len, "mc#%u", mci->mc_idx); in edac_mc_alloc_dimms()
293 for (layer = 0; layer < mci->n_layers; layer++) { in edac_mc_alloc_dimms()
295 edac_layer_name[mci->layers[layer].type], in edac_mc_alloc_dimms()
308 if (mci->layers[0].is_virt_csrow) { in edac_mc_alloc_dimms()
310 if (chn == mci->num_cschannel) { in edac_mc_alloc_dimms()
316 if (row == mci->nr_csrows) { in edac_mc_alloc_dimms()
323 for (layer = mci->n_layers - 1; layer >= 0; layer--) { in edac_mc_alloc_dimms()
325 if (pos[layer] < mci->layers[layer].size) in edac_mc_alloc_dimms()
339 struct mem_ctl_info *mci; in edac_mc_alloc() local
364 mci = kzalloc(sizeof(struct mem_ctl_info), GFP_KERNEL); in edac_mc_alloc()
365 if (!mci) in edac_mc_alloc()
368 mci->layers = kcalloc(n_layers, sizeof(struct edac_mc_layer), GFP_KERNEL); in edac_mc_alloc()
369 if (!mci->layers) in edac_mc_alloc()
372 mci->pvt_info = kzalloc(sz_pvt, GFP_KERNEL); in edac_mc_alloc()
373 if (!mci->pvt_info) in edac_mc_alloc()
376 mci->dev.release = mci_release; in edac_mc_alloc()
377 device_initialize(&mci->dev); in edac_mc_alloc()
380 mci->mc_idx = mc_num; in edac_mc_alloc()
381 mci->tot_dimms = tot_dimms; in edac_mc_alloc()
382 mci->n_layers = n_layers; in edac_mc_alloc()
383 memcpy(mci->layers, layers, sizeof(*layer) * n_layers); in edac_mc_alloc()
384 mci->nr_csrows = tot_csrows; in edac_mc_alloc()
385 mci->num_cschannel = tot_channels; in edac_mc_alloc()
386 mci->csbased = per_rank; in edac_mc_alloc()
388 if (edac_mc_alloc_csrows(mci)) in edac_mc_alloc()
391 if (edac_mc_alloc_dimms(mci)) in edac_mc_alloc()
394 mci->op_state = OP_ALLOC; in edac_mc_alloc()
396 return mci; in edac_mc_alloc()
399 _edac_mc_free(mci); in edac_mc_alloc()
405 void edac_mc_free(struct mem_ctl_info *mci) in edac_mc_free() argument
409 _edac_mc_free(mci); in edac_mc_free()
430 struct mem_ctl_info *mci; in __find_mci_by_dev() local
436 mci = list_entry(item, struct mem_ctl_info, link); in __find_mci_by_dev()
438 if (mci->pdev == dev) in __find_mci_by_dev()
439 return mci; in __find_mci_by_dev()
471 struct mem_ctl_info *mci = to_edac_mem_ctl_work(d_work); in edac_mc_workq_function() local
475 if (mci->op_state != OP_RUNNING_POLL) { in edac_mc_workq_function()
481 mci->edac_check(mci); in edac_mc_workq_function()
486 edac_queue_work(&mci->work, msecs_to_jiffies(edac_mc_get_poll_msec())); in edac_mc_workq_function()
497 struct mem_ctl_info *mci; in edac_mc_reset_delay_period() local
503 mci = list_entry(item, struct mem_ctl_info, link); in edac_mc_reset_delay_period()
505 if (mci->op_state == OP_RUNNING_POLL) in edac_mc_reset_delay_period()
506 edac_mod_work(&mci->work, value); in edac_mc_reset_delay_period()
521 static int add_mc_to_global_list(struct mem_ctl_info *mci) in add_mc_to_global_list() argument
528 p = __find_mci_by_dev(mci->pdev); in add_mc_to_global_list()
535 if (p->mc_idx >= mci->mc_idx) { in add_mc_to_global_list()
536 if (unlikely(p->mc_idx == mci->mc_idx)) in add_mc_to_global_list()
544 list_add_tail_rcu(&mci->link, insert_before); in add_mc_to_global_list()
550 edac_dev_name(mci), p->mod_name, p->ctl_name, p->mc_idx); in add_mc_to_global_list()
560 static int del_mc_from_global_list(struct mem_ctl_info *mci) in del_mc_from_global_list() argument
562 list_del_rcu(&mci->link); in del_mc_from_global_list()
568 INIT_LIST_HEAD(&mci->link); in del_mc_from_global_list()
575 struct mem_ctl_info *mci; in edac_mc_find() local
581 mci = list_entry(item, struct mem_ctl_info, link); in edac_mc_find()
582 if (mci->mc_idx == idx) in edac_mc_find()
586 mci = NULL; in edac_mc_find()
589 return mci; in edac_mc_find()
600 int edac_mc_add_mc_with_groups(struct mem_ctl_info *mci, in edac_mc_add_mc_with_groups() argument
608 edac_mc_dump_mci(mci); in edac_mc_add_mc_with_groups()
614 for (i = 0; i < mci->nr_csrows; i++) { in edac_mc_add_mc_with_groups()
615 struct csrow_info *csrow = mci->csrows[i]; in edac_mc_add_mc_with_groups()
629 mci_for_each_dimm(mci, dimm) in edac_mc_add_mc_with_groups()
635 if (edac_mc_owner && edac_mc_owner != mci->mod_name) { in edac_mc_add_mc_with_groups()
640 if (add_mc_to_global_list(mci)) in edac_mc_add_mc_with_groups()
644 mci->start_time = jiffies; in edac_mc_add_mc_with_groups()
646 mci->bus = edac_get_sysfs_subsys(); in edac_mc_add_mc_with_groups()
648 if (edac_create_sysfs_mci_device(mci, groups)) { in edac_mc_add_mc_with_groups()
649 edac_mc_printk(mci, KERN_WARNING, in edac_mc_add_mc_with_groups()
654 if (mci->edac_check) { in edac_mc_add_mc_with_groups()
655 mci->op_state = OP_RUNNING_POLL; in edac_mc_add_mc_with_groups()
657 INIT_DELAYED_WORK(&mci->work, edac_mc_workq_function); in edac_mc_add_mc_with_groups()
658 edac_queue_work(&mci->work, msecs_to_jiffies(edac_mc_get_poll_msec())); in edac_mc_add_mc_with_groups()
661 mci->op_state = OP_RUNNING_INTERRUPT; in edac_mc_add_mc_with_groups()
665 edac_mc_printk(mci, KERN_INFO, in edac_mc_add_mc_with_groups()
667 mci->mod_name, mci->ctl_name, mci->dev_name, in edac_mc_add_mc_with_groups()
668 edac_op_state_to_string(mci->op_state)); in edac_mc_add_mc_with_groups()
670 edac_mc_owner = mci->mod_name; in edac_mc_add_mc_with_groups()
676 del_mc_from_global_list(mci); in edac_mc_add_mc_with_groups()
686 struct mem_ctl_info *mci; in edac_mc_del_mc() local
693 mci = __find_mci_by_dev(dev); in edac_mc_del_mc()
694 if (mci == NULL) { in edac_mc_del_mc()
700 mci->op_state = OP_OFFLINE; in edac_mc_del_mc()
702 if (del_mc_from_global_list(mci)) in edac_mc_del_mc()
707 if (mci->edac_check) in edac_mc_del_mc()
708 edac_stop_work(&mci->work); in edac_mc_del_mc()
711 edac_remove_sysfs_mci_device(mci); in edac_mc_del_mc()
714 "Removed device %d for %s %s: DEV %s\n", mci->mc_idx, in edac_mc_del_mc()
715 mci->mod_name, mci->ctl_name, edac_dev_name(mci)); in edac_mc_del_mc()
717 return mci; in edac_mc_del_mc()
753 int edac_mc_find_csrow_by_page(struct mem_ctl_info *mci, unsigned long page) in edac_mc_find_csrow_by_page() argument
755 struct csrow_info **csrows = mci->csrows; in edac_mc_find_csrow_by_page()
758 edac_dbg(1, "MC%d: 0x%lx\n", mci->mc_idx, page); in edac_mc_find_csrow_by_page()
761 for (i = 0; i < mci->nr_csrows; i++) { in edac_mc_find_csrow_by_page()
772 mci->mc_idx, in edac_mc_find_csrow_by_page()
786 edac_mc_printk(mci, KERN_ERR, in edac_mc_find_csrow_by_page()
806 struct mem_ctl_info *mci = error_desc_to_mci(e); in edac_inc_ce_error() local
807 struct dimm_info *dimm = edac_get_dimm(mci, pos[0], pos[1], pos[2]); in edac_inc_ce_error()
809 mci->ce_mc += e->error_count; in edac_inc_ce_error()
814 mci->ce_noinfo_count += e->error_count; in edac_inc_ce_error()
820 struct mem_ctl_info *mci = error_desc_to_mci(e); in edac_inc_ue_error() local
821 struct dimm_info *dimm = edac_get_dimm(mci, pos[0], pos[1], pos[2]); in edac_inc_ue_error()
823 mci->ue_mc += e->error_count; in edac_inc_ue_error()
828 mci->ue_noinfo_count += e->error_count; in edac_inc_ue_error()
833 struct mem_ctl_info *mci = error_desc_to_mci(e); in edac_ce_error() local
837 edac_mc_printk(mci, KERN_WARNING, in edac_ce_error()
849 if (mci->scrub_mode == SCRUB_SW_SRC) { in edac_ce_error()
861 remapped_page = mci->ctl_page_to_phys ? in edac_ce_error()
862 mci->ctl_page_to_phys(mci, e->page_frame_number) : in edac_ce_error()
871 struct mem_ctl_info *mci = error_desc_to_mci(e); in edac_ue_error() local
874 edac_mc_printk(mci, KERN_WARNING, in edac_ue_error()
899 struct mem_ctl_info *mci = error_desc_to_mci(e); in edac_inc_csrow() local
909 mci->csrows[row]->ce_count += count; in edac_inc_csrow()
911 mci->csrows[row]->channels[chan]->ce_count += count; in edac_inc_csrow()
913 mci->csrows[row]->ue_count += count; in edac_inc_csrow()
919 struct mem_ctl_info *mci = error_desc_to_mci(e); in edac_raw_mc_handle_error() local
931 mci->mc_idx, e->top_layer, e->mid_layer, in edac_raw_mc_handle_error()
944 struct mem_ctl_info *mci, in edac_mc_handle_error() argument
960 struct edac_raw_error_desc *e = &mci->error_desc; in edac_mc_handle_error()
964 edac_dbg(3, "MC%d\n", mci->mc_idx); in edac_mc_handle_error()
985 for (i = 0; i < mci->n_layers; i++) { in edac_mc_handle_error()
986 if (pos[i] >= (int)mci->layers[i].size) { in edac_mc_handle_error()
988 edac_mc_printk(mci, KERN_ERR, in edac_mc_handle_error()
990 edac_layer_name[mci->layers[i].type], in edac_mc_handle_error()
991 pos[i], mci->layers[i].size); in edac_mc_handle_error()
1020 mci_for_each_dimm(mci, dimm) { in edac_mc_handle_error()
1055 mci->csbased ? "rank" : "dimm", in edac_mc_handle_error()
1080 for (i = 0; i < mci->n_layers; i++) { in edac_mc_handle_error()
1085 edac_layer_name[mci->layers[i].type], pos[i]); in edac_mc_handle_error()