Lines Matching full:desired

82  * @desired: desired memory for system operation
95 size_t desired; member
182 * pool is used to increase the reserve pool toward the desired entitlement
242 * Increase the reserve pool until the desired allocation is met. in vio_cmo_dealloc()
246 if (excess_freed && (vio_cmo.desired > vio_cmo.reserve.size)) { in vio_cmo_dealloc()
247 tmp = min(excess_freed, (vio_cmo.desired - vio_cmo.reserve.size)); in vio_cmo_dealloc()
369 * The list of devices is iterated through to recalculate the desired
401 cmo->desired = cmo->min; in vio_cmo_balance()
412 cmo->desired += (viodev->cmo.desired - VIO_CMO_MIN_ENT); in vio_cmo_balance()
427 if (viodev->cmo.desired <= level) { in vio_cmo_balance()
435 * desired level of entitlement for the device. in vio_cmo_balance()
438 chunk = min(chunk, (viodev->cmo.desired - in vio_cmo_balance()
613 * vio_cmo_set_dev_desired - Set desired entitlement for a device
616 * @desired: new desired entitlement level in bytes
619 * through sysfs. The desired entitlement level is changed and a balancing
622 void vio_cmo_set_dev_desired(struct vio_dev *viodev, size_t desired) in vio_cmo_set_dev_desired() argument
632 if (desired < VIO_CMO_MIN_ENT) in vio_cmo_set_dev_desired()
633 desired = VIO_CMO_MIN_ENT; in vio_cmo_set_dev_desired()
650 /* Increase/decrease in desired device entitlement */ in vio_cmo_set_dev_desired()
651 if (desired >= viodev->cmo.desired) { in vio_cmo_set_dev_desired()
653 vio_cmo.desired += desired - viodev->cmo.desired; in vio_cmo_set_dev_desired()
654 viodev->cmo.desired = desired; in vio_cmo_set_dev_desired()
656 /* Decrease bus and device values for desired entitlement */ in vio_cmo_set_dev_desired()
657 vio_cmo.desired -= viodev->cmo.desired - desired; in vio_cmo_set_dev_desired()
658 viodev->cmo.desired = desired; in vio_cmo_set_dev_desired()
660 * If less entitlement is desired than current entitlement, move in vio_cmo_set_dev_desired()
663 if (viodev->cmo.entitled > desired) { in vio_cmo_set_dev_desired()
664 vio_cmo.reserve.size -= viodev->cmo.entitled - desired; in vio_cmo_set_dev_desired()
665 vio_cmo.excess.size += viodev->cmo.entitled - desired; in vio_cmo_set_dev_desired()
673 max(viodev->cmo.allocated, desired); in vio_cmo_set_dev_desired()
674 viodev->cmo.entitled = desired; in vio_cmo_set_dev_desired()
725 /* Check that the driver is CMO enabled and get desired DMA */ in vio_cmo_bus_probe()
732 viodev->cmo.desired = in vio_cmo_bus_probe()
734 if (viodev->cmo.desired < VIO_CMO_MIN_ENT) in vio_cmo_bus_probe()
735 viodev->cmo.desired = VIO_CMO_MIN_ENT; in vio_cmo_bus_probe()
747 viodev->cmo.desired = 0; in vio_cmo_bus_probe()
760 /* Updated desired entitlement if device requires it */ in vio_cmo_bus_probe()
762 vio_cmo.desired += (viodev->cmo.desired - in vio_cmo_bus_probe()
788 vio_cmo.desired += viodev->cmo.desired; in vio_cmo_bus_probe()
838 * vio_cmo.desired in vio_cmo_bus_remove()
840 vio_cmo.desired -= (viodev->cmo.desired - VIO_CMO_MIN_ENT); in vio_cmo_bus_remove()
868 viodev->cmo.desired = VIO_CMO_MIN_ENT; in vio_cmo_bus_remove()
927 vio_cmo.desired = vio_cmo.reserve.size; in vio_cmo_bus_init()
970 viodev_cmo_rd_attr(desired);
1021 viobus_cmo_rd_attr(desired);
1066 void vio_cmo_set_dev_desired(struct vio_dev *viodev, size_t desired) {} in vio_cmo_set_dev_desired() argument