Lines Matching refs:subsys
117 static u32 nvmet_max_nsid(struct nvmet_subsys *subsys) in nvmet_max_nsid() argument
123 xa_for_each(&subsys->namespaces, idx, cur) in nvmet_max_nsid()
235 void nvmet_ns_changed(struct nvmet_subsys *subsys, u32 nsid) in nvmet_ns_changed() argument
239 lockdep_assert_held(&subsys->lock); in nvmet_ns_changed()
241 list_for_each_entry(ctrl, &subsys->ctrls, subsys_entry) { in nvmet_ns_changed()
251 void nvmet_send_ana_event(struct nvmet_subsys *subsys, in nvmet_send_ana_event() argument
256 mutex_lock(&subsys->lock); in nvmet_send_ana_event()
257 list_for_each_entry(ctrl, &subsys->ctrls, subsys_entry) { in nvmet_send_ana_event()
265 mutex_unlock(&subsys->lock); in nvmet_send_ana_event()
274 nvmet_send_ana_event(p->subsys, port); in nvmet_port_send_ana_event()
301 void nvmet_port_del_ctrls(struct nvmet_port *port, struct nvmet_subsys *subsys) in nvmet_port_del_ctrls() argument
305 mutex_lock(&subsys->lock); in nvmet_port_del_ctrls()
306 list_for_each_entry(ctrl, &subsys->ctrls, subsys_entry) { in nvmet_port_del_ctrls()
310 mutex_unlock(&subsys->lock); in nvmet_port_del_ctrls()
544 nvmet_ns_changed(ns->subsys, ns->nsid); in nvmet_ns_revalidate()
549 struct nvmet_subsys *subsys = ns->subsys; in nvmet_ns_enable() local
553 mutex_lock(&subsys->lock); in nvmet_ns_enable()
556 if (nvmet_is_passthru_subsys(subsys)) { in nvmet_ns_enable()
565 if (subsys->nr_namespaces == NVMET_MAX_NAMESPACES) in nvmet_ns_enable()
578 list_for_each_entry(ctrl, &subsys->ctrls, subsys_entry) in nvmet_ns_enable()
586 if (ns->nsid > subsys->max_nsid) in nvmet_ns_enable()
587 subsys->max_nsid = ns->nsid; in nvmet_ns_enable()
589 ret = xa_insert(&subsys->namespaces, ns->nsid, ns, GFP_KERNEL); in nvmet_ns_enable()
593 subsys->nr_namespaces++; in nvmet_ns_enable()
595 nvmet_ns_changed(subsys, ns->nsid); in nvmet_ns_enable()
599 mutex_unlock(&subsys->lock); in nvmet_ns_enable()
603 subsys->max_nsid = nvmet_max_nsid(subsys); in nvmet_ns_enable()
606 list_for_each_entry(ctrl, &subsys->ctrls, subsys_entry) in nvmet_ns_enable()
615 struct nvmet_subsys *subsys = ns->subsys; in nvmet_ns_disable() local
618 mutex_lock(&subsys->lock); in nvmet_ns_disable()
623 xa_erase(&ns->subsys->namespaces, ns->nsid); in nvmet_ns_disable()
624 if (ns->nsid == subsys->max_nsid) in nvmet_ns_disable()
625 subsys->max_nsid = nvmet_max_nsid(subsys); in nvmet_ns_disable()
627 list_for_each_entry(ctrl, &subsys->ctrls, subsys_entry) in nvmet_ns_disable()
630 mutex_unlock(&subsys->lock); in nvmet_ns_disable()
645 mutex_lock(&subsys->lock); in nvmet_ns_disable()
647 subsys->nr_namespaces--; in nvmet_ns_disable()
648 nvmet_ns_changed(subsys, ns->nsid); in nvmet_ns_disable()
651 mutex_unlock(&subsys->lock); in nvmet_ns_disable()
666 struct nvmet_ns *nvmet_ns_alloc(struct nvmet_subsys *subsys, u32 nsid) in nvmet_ns_alloc() argument
677 ns->subsys = subsys; in nvmet_ns_alloc()
1143 if (ctrl->subsys->type != NVME_NQN_DISC && in nvmet_start_ctrl()
1210 if (nvmet_is_passthru_subsys(ctrl->subsys)) in nvmet_init_cap()
1219 struct nvmet_subsys *subsys; in nvmet_ctrl_find_get() local
1221 subsys = nvmet_find_get_subsys(req->port, subsysnqn); in nvmet_ctrl_find_get()
1222 if (!subsys) { in nvmet_ctrl_find_get()
1229 mutex_lock(&subsys->lock); in nvmet_ctrl_find_get()
1230 list_for_each_entry(ctrl, &subsys->ctrls, subsys_entry) { in nvmet_ctrl_find_get()
1250 mutex_unlock(&subsys->lock); in nvmet_ctrl_find_get()
1251 nvmet_subsys_put(subsys); in nvmet_ctrl_find_get()
1272 bool nvmet_host_allowed(struct nvmet_subsys *subsys, const char *hostnqn) in nvmet_host_allowed() argument
1278 if (subsys->allow_any_host) in nvmet_host_allowed()
1281 if (subsys->type == NVME_NQN_DISC) /* allow all access to disc subsys */ in nvmet_host_allowed()
1284 list_for_each_entry(p, &subsys->hosts, entry) { in nvmet_host_allowed()
1306 xa_for_each(&ctrl->subsys->namespaces, idx, ns) in nvmet_setup_p2p_ns_map()
1336 struct nvmet_subsys *subsys; in nvmet_alloc_ctrl() local
1342 subsys = nvmet_find_get_subsys(req->port, subsysnqn); in nvmet_alloc_ctrl()
1343 if (!subsys) { in nvmet_alloc_ctrl()
1352 if (!nvmet_host_allowed(subsys, hostnqn)) { in nvmet_alloc_ctrl()
1381 ctrl->subsys = subsys; in nvmet_alloc_ctrl()
1390 ctrl->sqs = kcalloc(subsys->max_qid + 1, in nvmet_alloc_ctrl()
1396 if (subsys->cntlid_min > subsys->cntlid_max) in nvmet_alloc_ctrl()
1400 subsys->cntlid_min, subsys->cntlid_max, in nvmet_alloc_ctrl()
1414 if ((ctrl->subsys->type == NVME_NQN_DISC) && !kato) in nvmet_alloc_ctrl()
1425 mutex_lock(&subsys->lock); in nvmet_alloc_ctrl()
1426 list_add_tail(&ctrl->subsys_entry, &subsys->ctrls); in nvmet_alloc_ctrl()
1428 mutex_unlock(&subsys->lock); in nvmet_alloc_ctrl()
1440 nvmet_subsys_put(subsys); in nvmet_alloc_ctrl()
1448 struct nvmet_subsys *subsys = ctrl->subsys; in nvmet_ctrl_free() local
1450 mutex_lock(&subsys->lock); in nvmet_ctrl_free()
1453 mutex_unlock(&subsys->lock); in nvmet_ctrl_free()
1467 nvmet_subsys_put(subsys); in nvmet_ctrl_free()
1502 if (!strncmp(p->subsys->subsysnqn, subsysnqn, in nvmet_find_get_subsys()
1504 if (!kref_get_unless_zero(&p->subsys->ref)) in nvmet_find_get_subsys()
1507 return p->subsys; in nvmet_find_get_subsys()
1517 struct nvmet_subsys *subsys; in nvmet_subsys_alloc() local
1521 subsys = kzalloc(sizeof(*subsys), GFP_KERNEL); in nvmet_subsys_alloc()
1522 if (!subsys) in nvmet_subsys_alloc()
1525 subsys->ver = NVMET_DEFAULT_VS; in nvmet_subsys_alloc()
1528 bin2hex(subsys->serial, &serial, sizeof(serial)); in nvmet_subsys_alloc()
1530 subsys->model_number = kstrdup(NVMET_DEFAULT_CTRL_MODEL, GFP_KERNEL); in nvmet_subsys_alloc()
1531 if (!subsys->model_number) { in nvmet_subsys_alloc()
1538 subsys->max_qid = NVMET_NR_QUEUES; in nvmet_subsys_alloc()
1541 subsys->max_qid = 0; in nvmet_subsys_alloc()
1548 subsys->type = type; in nvmet_subsys_alloc()
1549 subsys->subsysnqn = kstrndup(subsysnqn, NVMF_NQN_SIZE, in nvmet_subsys_alloc()
1551 if (!subsys->subsysnqn) { in nvmet_subsys_alloc()
1555 subsys->cntlid_min = NVME_CNTLID_MIN; in nvmet_subsys_alloc()
1556 subsys->cntlid_max = NVME_CNTLID_MAX; in nvmet_subsys_alloc()
1557 kref_init(&subsys->ref); in nvmet_subsys_alloc()
1559 mutex_init(&subsys->lock); in nvmet_subsys_alloc()
1560 xa_init(&subsys->namespaces); in nvmet_subsys_alloc()
1561 INIT_LIST_HEAD(&subsys->ctrls); in nvmet_subsys_alloc()
1562 INIT_LIST_HEAD(&subsys->hosts); in nvmet_subsys_alloc()
1564 return subsys; in nvmet_subsys_alloc()
1567 kfree(subsys->model_number); in nvmet_subsys_alloc()
1569 kfree(subsys); in nvmet_subsys_alloc()
1575 struct nvmet_subsys *subsys = in nvmet_subsys_free() local
1578 WARN_ON_ONCE(!xa_empty(&subsys->namespaces)); in nvmet_subsys_free()
1580 xa_destroy(&subsys->namespaces); in nvmet_subsys_free()
1581 nvmet_passthru_subsys_free(subsys); in nvmet_subsys_free()
1583 kfree(subsys->subsysnqn); in nvmet_subsys_free()
1584 kfree(subsys->model_number); in nvmet_subsys_free()
1585 kfree(subsys); in nvmet_subsys_free()
1588 void nvmet_subsys_del_ctrls(struct nvmet_subsys *subsys) in nvmet_subsys_del_ctrls() argument
1592 mutex_lock(&subsys->lock); in nvmet_subsys_del_ctrls()
1593 list_for_each_entry(ctrl, &subsys->ctrls, subsys_entry) in nvmet_subsys_del_ctrls()
1595 mutex_unlock(&subsys->lock); in nvmet_subsys_del_ctrls()
1598 void nvmet_subsys_put(struct nvmet_subsys *subsys) in nvmet_subsys_put() argument
1600 kref_put(&subsys->ref, nvmet_subsys_free); in nvmet_subsys_put()