Home
last modified time | relevance | path

Searched refs:dev_to_node (Results 1 – 25 of 98) sorted by relevance

1234

/Linux-v5.4/include/linux/
Dasync.h90 return async_schedule_node(func, dev, dev_to_node(dev)); in async_schedule_dev()
112 return async_schedule_node_domain(func, dev, dev_to_node(dev), domain); in async_schedule_dev_domain()
/Linux-v5.4/drivers/crypto/qat/qat_c3xxx/
Dadf_drv.c137 if (num_possible_nodes() > 1 && dev_to_node(&pdev->dev) < 0) { in adf_probe()
146 dev_to_node(&pdev->dev)); in adf_probe()
165 dev_to_node(&pdev->dev)); in adf_probe()
/Linux-v5.4/drivers/crypto/qat/qat_c62x/
Dadf_drv.c137 if (num_possible_nodes() > 1 && dev_to_node(&pdev->dev) < 0) { in adf_probe()
146 dev_to_node(&pdev->dev)); in adf_probe()
165 dev_to_node(&pdev->dev)); in adf_probe()
/Linux-v5.4/drivers/crypto/qat/qat_dh895xcc/
Dadf_drv.c137 if (num_possible_nodes() > 1 && dev_to_node(&pdev->dev) < 0) { in adf_probe()
146 dev_to_node(&pdev->dev)); in adf_probe()
165 dev_to_node(&pdev->dev)); in adf_probe()
/Linux-v5.4/drivers/crypto/virtio/
Dvirtio_crypto_mgr.c194 if ((node == dev_to_node(&tmp_dev->vdev->dev) || in virtcrypto_get_dev_node()
195 dev_to_node(&tmp_dev->vdev->dev) < 0) && in virtcrypto_get_dev_node()
Dvirtio_crypto_algs.c373 dev_to_node(&vcrypto->vdev->dev)); in __virtio_crypto_ablkcipher_do_req()
378 dev_to_node(&vcrypto->vdev->dev)); in __virtio_crypto_ablkcipher_do_req()
434 dev_to_node(&vcrypto->vdev->dev)); in __virtio_crypto_ablkcipher_do_req()
Dvirtio_crypto_core.c311 if (num_possible_nodes() > 1 && dev_to_node(&vdev->dev) < 0) { in virtcrypto_probe()
322 dev_to_node(&vdev->dev)); in virtcrypto_probe()
/Linux-v5.4/drivers/crypto/qat/qat_common/
Dqat_crypto.c104 if ((node == dev_to_node(&GET_DEV(tmp_dev)) || in qat_crypto_get_instance_node()
105 dev_to_node(&GET_DEV(tmp_dev)) < 0) && in qat_crypto_get_instance_node()
262 dev_to_node(&GET_DEV(accel_dev))); in qat_crypto_create_instances()
Dadf_transport.c418 dev_to_node(&GET_DEV(accel_dev))); in adf_init_bank()
469 dev_to_node(&GET_DEV(accel_dev))); in adf_init_etr_data()
476 dev_to_node(&GET_DEV(accel_dev))); in adf_init_etr_data()
Dadf_admin.c244 dev_to_node(&GET_DEV(accel_dev))); in adf_init_admin_comms()
Dadf_isr.c242 GFP_KERNEL, dev_to_node(&GET_DEV(accel_dev))); in adf_isr_alloc_msix_entry_table()
/Linux-v5.4/drivers/usb/host/
Dxhci-mem.c38 seg = kzalloc_node(sizeof(*seg), flags, dev_to_node(dev)); in xhci_segment_alloc()
50 dev_to_node(dev)); in xhci_segment_alloc()
370 ring = kzalloc_node(sizeof(*ring), flags, dev_to_node(dev)); in xhci_ring_alloc()
469 ctx = kzalloc_node(sizeof(*ctx), flags, dev_to_node(dev)); in xhci_alloc_container_ctx()
634 dev_to_node(dev)); in xhci_alloc_stream_info()
644 dev_to_node(dev)); in xhci_alloc_stream_info()
851 dev_to_node(dev)); in xhci_alloc_tt_info()
1657 dev_to_node(dev)); in scratchpad_alloc()
1668 flags, dev_to_node(dev)); in scratchpad_alloc()
1738 command = kzalloc_node(sizeof(*command), mem_flags, dev_to_node(dev)); in xhci_alloc_command()
[all …]
/Linux-v5.4/drivers/base/test/
Dtest_async_driver_probe.c47 if (dev_to_node(dev) != numa_node_id()) { in test_probe()
49 dev_to_node(dev), numa_node_id()); in test_probe()
/Linux-v5.4/drivers/crypto/qat/qat_c62xvf/
Dadf_drv.c140 dev_to_node(&pdev->dev)); in adf_probe()
160 dev_to_node(&pdev->dev)); in adf_probe()
/Linux-v5.4/drivers/crypto/qat/qat_c3xxxvf/
Dadf_drv.c140 dev_to_node(&pdev->dev)); in adf_probe()
160 dev_to_node(&pdev->dev)); in adf_probe()
/Linux-v5.4/drivers/crypto/qat/qat_dh895xccvf/
Dadf_drv.c140 dev_to_node(&pdev->dev)); in adf_probe()
160 dev_to_node(&pdev->dev)); in adf_probe()
/Linux-v5.4/drivers/nvdimm/
Dof_pmem.c70 ndr_desc.numa_node = dev_to_node(&pdev->dev); in of_pmem_region_probe()
Dvirtio_pmem.c35 int nid = dev_to_node(&vdev->dev); in virtio_pmem_probe()
Dpmem.c357 int nid = dev_to_node(dev), fua; in pmem_attach_disk()
401 q = blk_alloc_queue_node(GFP_KERNEL, dev_to_node(dev)); in pmem_attach_disk()
/Linux-v5.4/kernel/
Diomem.c145 dev_to_node(dev)); in devm_memremap()
/Linux-v5.4/drivers/perf/
Dthunderx2_pmu.c651 tx2_pmu->node = dev_to_node(dev); in tx2_uncore_pmu_init_dev()
792 dev_info(dev, "node%d: pmu uncore registered\n", dev_to_node(dev)); in tx2_uncore_probe()
803 if (tx2_pmu->node == dev_to_node(dev)) { in tx2_uncore_remove()
/Linux-v5.4/arch/arm64/kernel/
Dpci.c61 return dev_to_node(&bus->dev); in pcibus_to_node()
/Linux-v5.4/arch/powerpc/kernel/
Ddma-iommu.c46 dev_to_node(dev)); in dma_iommu_alloc_coherent()
/Linux-v5.4/arch/x86/pci/
Dintel_mid_pci.c258 ioapic_set_alloc_attr(&info, dev_to_node(&dev->dev), 1, polarity); in intel_mid_pci_irq_enable()
/Linux-v5.4/drivers/pci/
Dp2pdma.c123 p2p->pool = gen_pool_create(PAGE_SHIFT, dev_to_node(&pdev->dev)); in pci_p2pdma_setup()
205 resource_size(&pgmap->res), dev_to_node(&pdev->dev), in pci_p2pdma_add_resource()

1234