Lines Matching refs:dev_dax
62 struct dev_dax *dev_dax = to_dev_dax(dev); in dax_match_type() local
64 if (dev_dax->region->res.flags & IORESOURCE_DAX_KMEM) in dax_match_type()
172 bool static_dev_dax(struct dev_dax *dev_dax) in static_dev_dax() argument
174 return is_static(dev_dax->region); in static_dev_dax()
178 static u64 dev_dax_size(struct dev_dax *dev_dax) in dev_dax_size() argument
183 device_lock_assert(&dev_dax->dev); in dev_dax_size()
185 for (i = 0; i < dev_dax->nr_range; i++) in dev_dax_size()
186 size += range_len(&dev_dax->ranges[i].range); in dev_dax_size()
194 struct dev_dax *dev_dax = to_dev_dax(dev); in dax_bus_probe() local
195 struct dax_region *dax_region = dev_dax->region; in dax_bus_probe()
198 if (dev_dax_size(dev_dax) == 0 || dev_dax->id < 0) in dax_bus_probe()
201 rc = dax_drv->probe(dev_dax); in dax_bus_probe()
219 struct dev_dax *dev_dax = to_dev_dax(dev); in dax_bus_remove() local
222 dax_drv->remove(dev_dax); in dax_bus_remove()
371 struct dev_dax *dev_dax = devm_create_dev_dax(&data); in create_store() local
373 if (IS_ERR(dev_dax)) in create_store()
374 rc = PTR_ERR(dev_dax); in create_store()
384 dax_region->seed = &dev_dax->dev; in create_store()
385 dax_region->youngest = &dev_dax->dev; in create_store()
395 void kill_dev_dax(struct dev_dax *dev_dax) in kill_dev_dax() argument
397 struct dax_device *dax_dev = dev_dax->dax_dev; in kill_dev_dax()
408 if (!static_dev_dax(dev_dax)) in kill_dev_dax()
409 dev_dax->pgmap = NULL; in kill_dev_dax()
413 static void trim_dev_dax_range(struct dev_dax *dev_dax) in trim_dev_dax_range() argument
415 int i = dev_dax->nr_range - 1; in trim_dev_dax_range()
416 struct range *range = &dev_dax->ranges[i].range; in trim_dev_dax_range()
417 struct dax_region *dax_region = dev_dax->region; in trim_dev_dax_range()
420 dev_dbg(&dev_dax->dev, "delete range[%d]: %#llx:%#llx\n", i, in trim_dev_dax_range()
425 if (--dev_dax->nr_range == 0) { in trim_dev_dax_range()
426 kfree(dev_dax->ranges); in trim_dev_dax_range()
427 dev_dax->ranges = NULL; in trim_dev_dax_range()
431 static void free_dev_dax_ranges(struct dev_dax *dev_dax) in free_dev_dax_ranges() argument
433 while (dev_dax->nr_range) in free_dev_dax_ranges()
434 trim_dev_dax_range(dev_dax); in free_dev_dax_ranges()
439 struct dev_dax *dev_dax = to_dev_dax(dev); in unregister_dev_dax() local
443 kill_dev_dax(dev_dax); in unregister_dev_dax()
445 free_dev_dax_ranges(dev_dax); in unregister_dev_dax()
463 static int __free_dev_dax_id(struct dev_dax *dev_dax) in __free_dev_dax_id() argument
465 struct device *dev = &dev_dax->dev; in __free_dev_dax_id()
467 int rc = dev_dax->id; in __free_dev_dax_id()
471 if (!dev_dax->dyn_id || dev_dax->id < 0) in __free_dev_dax_id()
473 dax_region = dev_dax->region; in __free_dev_dax_id()
474 ida_free(&dax_region->ida, dev_dax->id); in __free_dev_dax_id()
476 dev_dax->id = -1; in __free_dev_dax_id()
480 static int free_dev_dax_id(struct dev_dax *dev_dax) in free_dev_dax_id() argument
482 struct device *dev = &dev_dax->dev; in free_dev_dax_id()
486 rc = __free_dev_dax_id(dev_dax); in free_dev_dax_id()
491 static int alloc_dev_dax_id(struct dev_dax *dev_dax) in alloc_dev_dax_id() argument
493 struct dax_region *dax_region = dev_dax->region; in alloc_dev_dax_id()
500 dev_dax->dyn_id = true; in alloc_dev_dax_id()
501 dev_dax->id = id; in alloc_dev_dax_id()
509 struct dev_dax *dev_dax; in delete_store() local
523 dev_dax = to_dev_dax(victim); in delete_store()
524 if (victim->driver || dev_dax_size(dev_dax)) in delete_store()
533 if (dev_dax->id > 0) { in delete_store()
534 do_del = __free_dev_dax_id(dev_dax) >= 0; in delete_store()
653 struct dev_dax *dev_dax = to_dev_dax(parent); in dax_mapping_release() local
655 ida_free(&dev_dax->ida, mapping->id); in dax_mapping_release()
664 struct dev_dax *dev_dax = to_dev_dax(dev->parent); in unregister_dax_mapping() local
665 struct dax_region *dax_region = dev_dax->region; in unregister_dax_mapping()
671 dev_dax->ranges[mapping->range_id].mapping = NULL; in unregister_dax_mapping()
680 struct dev_dax *dev_dax = to_dev_dax(dev->parent); in get_dax_range() local
681 struct dax_region *dax_region = dev_dax->region; in get_dax_range()
689 return &dev_dax->ranges[mapping->range_id]; in get_dax_range()
695 struct dev_dax *dev_dax = to_dev_dax(mapping->dev.parent); in put_dax_range() local
696 struct dax_region *dax_region = dev_dax->region; in put_dax_range()
770 static int devm_register_dax_mapping(struct dev_dax *dev_dax, int range_id) in devm_register_dax_mapping() argument
772 struct dax_region *dax_region = dev_dax->region; in devm_register_dax_mapping()
779 if (dev_WARN_ONCE(&dev_dax->dev, !dax_region->dev->driver, in devm_register_dax_mapping()
787 mapping->id = ida_alloc(&dev_dax->ida, GFP_KERNEL); in devm_register_dax_mapping()
792 dev_dax->ranges[range_id].mapping = mapping; in devm_register_dax_mapping()
795 dev->parent = &dev_dax->dev; in devm_register_dax_mapping()
812 static int alloc_dev_dax_range(struct dev_dax *dev_dax, u64 start, in alloc_dev_dax_range() argument
815 struct dax_region *dax_region = dev_dax->region; in alloc_dev_dax_range()
817 struct device *dev = &dev_dax->dev; in alloc_dev_dax_range()
827 if (dev_WARN_ONCE(dev, dev_dax->nr_range, in alloc_dev_dax_range()
838 ranges = krealloc(dev_dax->ranges, sizeof(*ranges) in alloc_dev_dax_range()
839 * (dev_dax->nr_range + 1), GFP_KERNEL); in alloc_dev_dax_range()
845 for (i = 0; i < dev_dax->nr_range; i++) in alloc_dev_dax_range()
847 dev_dax->ranges = ranges; in alloc_dev_dax_range()
848 ranges[dev_dax->nr_range++] = (struct dev_dax_range) { in alloc_dev_dax_range()
856 dev_dbg(dev, "alloc range[%d]: %pa:%pa\n", dev_dax->nr_range - 1, in alloc_dev_dax_range()
863 if (!device_is_registered(&dev_dax->dev)) in alloc_dev_dax_range()
866 rc = devm_register_dax_mapping(dev_dax, dev_dax->nr_range - 1); in alloc_dev_dax_range()
868 trim_dev_dax_range(dev_dax); in alloc_dev_dax_range()
873 static int adjust_dev_dax_range(struct dev_dax *dev_dax, struct resource *res, resource_size_t size) in adjust_dev_dax_range() argument
875 int last_range = dev_dax->nr_range - 1; in adjust_dev_dax_range()
876 struct dev_dax_range *dax_range = &dev_dax->ranges[last_range]; in adjust_dev_dax_range()
877 struct dax_region *dax_region = dev_dax->region; in adjust_dev_dax_range()
880 struct device *dev = &dev_dax->dev; in adjust_dev_dax_range()
907 struct dev_dax *dev_dax = to_dev_dax(dev); in size_show() local
911 size = dev_dax_size(dev_dax); in size_show()
917 static bool alloc_is_aligned(struct dev_dax *dev_dax, resource_size_t size) in alloc_is_aligned() argument
923 return IS_ALIGNED(size, max_t(unsigned long, dev_dax->align, memremap_compat_align())); in alloc_is_aligned()
926 static int dev_dax_shrink(struct dev_dax *dev_dax, resource_size_t size) in dev_dax_shrink() argument
928 resource_size_t to_shrink = dev_dax_size(dev_dax) - size; in dev_dax_shrink()
929 struct dax_region *dax_region = dev_dax->region; in dev_dax_shrink()
930 struct device *dev = &dev_dax->dev; in dev_dax_shrink()
933 for (i = dev_dax->nr_range - 1; i >= 0; i--) { in dev_dax_shrink()
934 struct range *range = &dev_dax->ranges[i].range; in dev_dax_shrink()
935 struct dax_mapping *mapping = dev_dax->ranges[i].mapping; in dev_dax_shrink()
943 trim_dev_dax_range(dev_dax); in dev_dax_shrink()
957 if (dev_WARN_ONCE(dev, !adjust || i != dev_dax->nr_range - 1, in dev_dax_shrink()
960 return adjust_dev_dax_range(dev_dax, adjust, range_len(range) in dev_dax_shrink()
970 static bool adjust_ok(struct dev_dax *dev_dax, struct resource *res) in adjust_ok() argument
975 if (dev_dax->nr_range == 0) in adjust_ok()
977 if (strcmp(res->name, dev_name(&dev_dax->dev)) != 0) in adjust_ok()
979 last = &dev_dax->ranges[dev_dax->nr_range - 1]; in adjust_ok()
982 for (i = 0; i < dev_dax->nr_range - 1; i++) { in adjust_ok()
983 struct dev_dax_range *dax_range = &dev_dax->ranges[i]; in adjust_ok()
993 struct dev_dax *dev_dax, resource_size_t size) in dev_dax_resize() argument
996 resource_size_t dev_size = dev_dax_size(dev_dax); in dev_dax_resize()
998 struct device *dev = &dev_dax->dev; in dev_dax_resize()
1010 return dev_dax_shrink(dev_dax, size); in dev_dax_resize()
1013 if (dev_WARN_ONCE(dev, !alloc_is_aligned(dev_dax, to_alloc), in dev_dax_resize()
1025 return alloc_dev_dax_range(dev_dax, dax_region->res.start, to_alloc); in dev_dax_resize()
1034 rc = alloc_dev_dax_range(dev_dax, dax_region->res.start, alloc); in dev_dax_resize()
1050 if (adjust_ok(dev_dax, res)) { in dev_dax_resize()
1051 rc = adjust_dev_dax_range(dev_dax, res, resource_size(res) + alloc); in dev_dax_resize()
1054 rc = alloc_dev_dax_range(dev_dax, res->end + 1, alloc); in dev_dax_resize()
1070 struct dev_dax *dev_dax = to_dev_dax(dev); in size_store() local
1071 struct dax_region *dax_region = dev_dax->region; in size_store()
1077 if (!alloc_is_aligned(dev_dax, val)) { in size_store()
1088 rc = dev_dax_resize(dax_region, dev_dax, val); in size_store()
1129 struct dev_dax *dev_dax = to_dev_dax(dev); in mapping_store() local
1130 struct dax_region *dax_region = dev_dax->region; in mapping_store()
1148 if (alloc_is_aligned(dev_dax, to_alloc)) in mapping_store()
1149 rc = alloc_dev_dax_range(dev_dax, r.start, to_alloc); in mapping_store()
1160 struct dev_dax *dev_dax = to_dev_dax(dev); in align_show() local
1162 return sprintf(buf, "%d\n", dev_dax->align); in align_show()
1165 static ssize_t dev_dax_validate_align(struct dev_dax *dev_dax) in dev_dax_validate_align() argument
1167 struct device *dev = &dev_dax->dev; in dev_dax_validate_align()
1170 for (i = 0; i < dev_dax->nr_range; i++) { in dev_dax_validate_align()
1171 size_t len = range_len(&dev_dax->ranges[i].range); in dev_dax_validate_align()
1173 if (!alloc_is_aligned(dev_dax, len)) { in dev_dax_validate_align()
1175 __func__, dev_dax->align, i); in dev_dax_validate_align()
1186 struct dev_dax *dev_dax = to_dev_dax(dev); in align_store() local
1187 struct dax_region *dax_region = dev_dax->region; in align_store()
1210 align_save = dev_dax->align; in align_store()
1211 dev_dax->align = val; in align_store()
1212 rc = dev_dax_validate_align(dev_dax); in align_store()
1214 dev_dax->align = align_save; in align_store()
1222 static int dev_dax_target_node(struct dev_dax *dev_dax) in dev_dax_target_node() argument
1224 struct dax_region *dax_region = dev_dax->region; in dev_dax_target_node()
1232 struct dev_dax *dev_dax = to_dev_dax(dev); in target_node_show() local
1234 return sprintf(buf, "%d\n", dev_dax_target_node(dev_dax)); in target_node_show()
1241 struct dev_dax *dev_dax = to_dev_dax(dev); in resource_show() local
1242 struct dax_region *dax_region = dev_dax->region; in resource_show()
1245 if (dev_dax->nr_range < 1) in resource_show()
1248 start = dev_dax->ranges[0].range.start; in resource_show()
1275 struct dev_dax *dev_dax = to_dev_dax(dev); in dev_dax_visible() local
1276 struct dax_region *dax_region = dev_dax->region; in dev_dax_visible()
1278 if (a == &dev_attr_target_node.attr && dev_dax_target_node(dev_dax) < 0) in dev_dax_visible()
1313 struct dev_dax *dev_dax = to_dev_dax(dev); in dev_dax_release() local
1314 struct dax_device *dax_dev = dev_dax->dax_dev; in dev_dax_release()
1317 free_dev_dax_id(dev_dax); in dev_dax_release()
1318 kfree(dev_dax->pgmap); in dev_dax_release()
1319 kfree(dev_dax); in dev_dax_release()
1327 struct dev_dax *devm_create_dev_dax(struct dev_dax_data *data) in devm_create_dev_dax()
1332 struct dev_dax *dev_dax; in devm_create_dev_dax() local
1337 dev_dax = kzalloc(sizeof(*dev_dax), GFP_KERNEL); in devm_create_dev_dax()
1338 if (!dev_dax) in devm_create_dev_dax()
1341 dev_dax->region = dax_region; in devm_create_dev_dax()
1349 dev_dax->id = data->id; in devm_create_dev_dax()
1357 rc = alloc_dev_dax_id(dev_dax); in devm_create_dev_dax()
1362 dev = &dev_dax->dev; in devm_create_dev_dax()
1364 dev_set_name(dev, "dax%d.%d", dax_region->id, dev_dax->id); in devm_create_dev_dax()
1366 rc = alloc_dev_dax_range(dev_dax, dax_region->res.start, data->size); in devm_create_dev_dax()
1374 dev_dax->pgmap = kmemdup(data->pgmap, in devm_create_dev_dax()
1376 if (!dev_dax->pgmap) { in devm_create_dev_dax()
1386 dax_dev = alloc_dax(dev_dax, NULL); in devm_create_dev_dax()
1398 dev_dax->dax_dev = dax_dev; in devm_create_dev_dax()
1399 dev_dax->target_node = dax_region->target_node; in devm_create_dev_dax()
1400 dev_dax->align = dax_region->align; in devm_create_dev_dax()
1401 ida_init(&dev_dax->ida); in devm_create_dev_dax()
1411 kill_dev_dax(dev_dax); in devm_create_dev_dax()
1421 if (dev_dax->nr_range && range_len(&dev_dax->ranges[0].range)) { in devm_create_dev_dax()
1422 rc = devm_register_dax_mapping(dev_dax, 0); in devm_create_dev_dax()
1427 return dev_dax; in devm_create_dev_dax()
1430 kfree(dev_dax->pgmap); in devm_create_dev_dax()
1432 free_dev_dax_ranges(dev_dax); in devm_create_dev_dax()
1434 free_dev_dax_id(dev_dax); in devm_create_dev_dax()
1436 kfree(dev_dax); in devm_create_dev_dax()