Lines Matching full:desired

83  * @desired: desired memory for system operation
96 size_t desired; member
183 * pool is used to increase the reserve pool toward the desired entitlement
243 * Increase the reserve pool until the desired allocation is met. in vio_cmo_dealloc()
247 if (excess_freed && (vio_cmo.desired > vio_cmo.reserve.size)) { in vio_cmo_dealloc()
248 tmp = min(excess_freed, (vio_cmo.desired - vio_cmo.reserve.size)); in vio_cmo_dealloc()
370 * The list of devices is iterated through to recalculate the desired
402 cmo->desired = cmo->min; in vio_cmo_balance()
413 cmo->desired += (viodev->cmo.desired - VIO_CMO_MIN_ENT); in vio_cmo_balance()
428 if (viodev->cmo.desired <= level) { in vio_cmo_balance()
436 * desired level of entitlement for the device. in vio_cmo_balance()
439 chunk = min(chunk, (viodev->cmo.desired - in vio_cmo_balance()
616 * vio_cmo_set_dev_desired - Set desired entitlement for a device
619 * @desired: new desired entitlement level in bytes
622 * through sysfs. The desired entitlement level is changed and a balancing
625 void vio_cmo_set_dev_desired(struct vio_dev *viodev, size_t desired) in vio_cmo_set_dev_desired() argument
635 if (desired < VIO_CMO_MIN_ENT) in vio_cmo_set_dev_desired()
636 desired = VIO_CMO_MIN_ENT; in vio_cmo_set_dev_desired()
653 /* Increase/decrease in desired device entitlement */ in vio_cmo_set_dev_desired()
654 if (desired >= viodev->cmo.desired) { in vio_cmo_set_dev_desired()
656 vio_cmo.desired += desired - viodev->cmo.desired; in vio_cmo_set_dev_desired()
657 viodev->cmo.desired = desired; in vio_cmo_set_dev_desired()
659 /* Decrease bus and device values for desired entitlement */ in vio_cmo_set_dev_desired()
660 vio_cmo.desired -= viodev->cmo.desired - desired; in vio_cmo_set_dev_desired()
661 viodev->cmo.desired = desired; in vio_cmo_set_dev_desired()
663 * If less entitlement is desired than current entitlement, move in vio_cmo_set_dev_desired()
666 if (viodev->cmo.entitled > desired) { in vio_cmo_set_dev_desired()
667 vio_cmo.reserve.size -= viodev->cmo.entitled - desired; in vio_cmo_set_dev_desired()
668 vio_cmo.excess.size += viodev->cmo.entitled - desired; in vio_cmo_set_dev_desired()
676 max(viodev->cmo.allocated, desired); in vio_cmo_set_dev_desired()
677 viodev->cmo.entitled = desired; in vio_cmo_set_dev_desired()
728 /* Check that the driver is CMO enabled and get desired DMA */ in vio_cmo_bus_probe()
735 viodev->cmo.desired = in vio_cmo_bus_probe()
737 if (viodev->cmo.desired < VIO_CMO_MIN_ENT) in vio_cmo_bus_probe()
738 viodev->cmo.desired = VIO_CMO_MIN_ENT; in vio_cmo_bus_probe()
750 viodev->cmo.desired = 0; in vio_cmo_bus_probe()
763 /* Updated desired entitlement if device requires it */ in vio_cmo_bus_probe()
765 vio_cmo.desired += (viodev->cmo.desired - in vio_cmo_bus_probe()
791 vio_cmo.desired += viodev->cmo.desired; in vio_cmo_bus_probe()
841 * vio_cmo.desired in vio_cmo_bus_remove()
843 vio_cmo.desired -= (viodev->cmo.desired - VIO_CMO_MIN_ENT); in vio_cmo_bus_remove()
871 viodev->cmo.desired = VIO_CMO_MIN_ENT; in vio_cmo_bus_remove()
930 vio_cmo.desired = vio_cmo.reserve.size; in vio_cmo_bus_init()
973 viodev_cmo_rd_attr(desired);
1024 viobus_cmo_rd_attr(desired);
1069 void vio_cmo_set_dev_desired(struct vio_dev *viodev, size_t desired) {} in vio_cmo_set_dev_desired() argument