/linux/drivers/acpi/acpica/ |
H A D | utdelete.c | 368 u16 new_count = 0; in acpi_ut_update_ref_count() local 390 new_count = original_count + 1; in acpi_ut_update_ref_count() 391 object->common.reference_count = new_count; in acpi_ut_update_ref_count() 406 new_count)); in acpi_ut_update_ref_count() 415 new_count = original_count - 1; in acpi_ut_update_ref_count() 416 object->common.reference_count = new_count; in acpi_ut_update_ref_count() 431 object->common.type, new_count)); in acpi_ut_update_ref_count() 435 if (new_count == 0) { in acpi_ut_update_ref_count() 453 if (new_count > ACPI_MAX_REFERENCE_COUNT) { in acpi_ut_update_ref_count() 456 new_count, objec in acpi_ut_update_ref_count() [all...] |
H A D | nsrepair.c | 447 u32 new_count; in acpi_ns_remove_null_elements() local 476 new_count = count; in acpi_ns_remove_null_elements() 485 new_count--; in acpi_ns_remove_null_elements() 496 if (new_count < count) { in acpi_ns_remove_null_elements() 499 info->full_pathname, (count - new_count))); in acpi_ns_remove_null_elements() 504 obj_desc->package.count = new_count; in acpi_ns_remove_null_elements()
|
H A D | nsrepair2.c | 922 u32 new_count; in acpi_ns_remove_element() local 928 new_count = count - 1; in acpi_ns_remove_element() 950 obj_desc->package.count = new_count; in acpi_ns_remove_element()
|
/linux/fs/ocfs2/ |
H A D | blockcheck.c | 283 u64 new_count; in ocfs2_blockcheck_inc_check() local 290 new_count = stats->b_check_count; in ocfs2_blockcheck_inc_check() 293 if (!new_count) in ocfs2_blockcheck_inc_check() 299 u64 new_count; in ocfs2_blockcheck_inc_failure() local 306 new_count = stats->b_failure_count; in ocfs2_blockcheck_inc_failure() 309 if (!new_count) in ocfs2_blockcheck_inc_failure() 315 u64 new_count; in ocfs2_blockcheck_inc_recover() local 322 new_count = stats->b_recover_count; in ocfs2_blockcheck_inc_recover() 325 if (!new_count) in ocfs2_blockcheck_inc_recover()
|
/linux/lib/ |
H A D | parman.c | 61 unsigned long new_count = parman->limit_count + in parman_enlarge() local 65 err = parman->ops->resize(parman->priv, new_count); in parman_enlarge() 68 parman->limit_count = new_count; in parman_enlarge() 74 unsigned long new_count = parman->limit_count - in parman_shrink() local 78 if (new_count < parman->ops->base_count) in parman_shrink() 80 err = parman->ops->resize(parman->priv, new_count); in parman_shrink() 83 parman->limit_count = new_count; in parman_shrink()
|
H A D | test_parman.c | 92 static int test_parman_resize(void *priv, unsigned long new_count) in test_parman_resize() argument 99 ITEM_PTRS_SIZE(new_count), GFP_KERNEL); in test_parman_resize() 100 if (new_count == 0) in test_parman_resize() 105 if (new_count > old_count) in test_parman_resize() 107 ITEM_PTRS_SIZE(new_count - old_count)); in test_parman_resize() 109 test_parman->prio_array_limit = new_count; in test_parman_resize()
|
/linux/io_uring/ |
H A D | register.c | 264 __u32 new_count[2]; in io_register_iowq_max_workers() local 267 if (copy_from_user(new_count, arg, sizeof(new_count))) in io_register_iowq_max_workers() 269 for (i = 0; i < ARRAY_SIZE(new_count); i++) in io_register_iowq_max_workers() 270 if (new_count[i] > INT_MAX) in io_register_iowq_max_workers() 295 BUILD_BUG_ON(sizeof(new_count) != sizeof(ctx->iowq_limits)); in io_register_iowq_max_workers() 297 for (i = 0; i < ARRAY_SIZE(new_count); i++) in io_register_iowq_max_workers() 298 if (new_count[i]) in io_register_iowq_max_workers() 299 ctx->iowq_limits[i] = new_count[i]; in io_register_iowq_max_workers() 303 ret = io_wq_max_workers(tctx->io_wq, new_count); in io_register_iowq_max_workers() [all...] |
H A D | io-wq.h | 49 int io_wq_max_workers(struct io_wq *wq, int *new_count);
|
H A D | io-wq.c | 1422 * Set max number of unbounded workers, returns old value. If new_count is 0, 1425 int io_wq_max_workers(struct io_wq *wq, int *new_count) in io_wq_max_workers() argument 1436 if (new_count[i] > task_rlimit(current, RLIMIT_NPROC)) in io_wq_max_workers() 1437 new_count[i] = task_rlimit(current, RLIMIT_NPROC); in io_wq_max_workers() 1449 if (new_count[i]) in io_wq_max_workers() 1450 acct->max_workers = new_count[i]; in io_wq_max_workers() 1456 new_count[i] = prev[i]; in io_wq_max_workers()
|
/linux/arch/arm/mm/ |
H A D | cache-l2x0-pmu.c | 120 u64 prev_count, new_count, mask; in l2x0_pmu_event_read() local 124 new_count = l2x0_pmu_counter_read(hw->idx); in l2x0_pmu_event_read() 125 } while (local64_xchg(&hw->prev_count, new_count) != prev_count); in l2x0_pmu_event_read() 128 local64_add((new_count - prev_count) & mask, &event->count); in l2x0_pmu_event_read() 130 warn_if_saturated(new_count); in l2x0_pmu_event_read()
|
/linux/drivers/net/ethernet/mellanox/mlxsw/ |
H A D | spectrum_acl_ctcam.c | 122 unsigned long new_count) in mlxsw_sp_acl_ctcam_region_parman_resize() argument 130 if (new_count > max_tcam_rules) in mlxsw_sp_acl_ctcam_region_parman_resize() 132 return mlxsw_sp_acl_ctcam_region_resize(mlxsw_sp, region, new_count); in mlxsw_sp_acl_ctcam_region_parman_resize()
|
H A D | spectrum1_mr_tcam.c | 199 unsigned long new_count) in mlxsw_sp1_mr_tcam_region_parman_resize() argument 207 if (new_count > max_tcam_rules) in mlxsw_sp1_mr_tcam_region_parman_resize() 210 mr_tcam_region->rtar_key_type, new_count); in mlxsw_sp1_mr_tcam_region_parman_resize()
|
/linux/drivers/perf/ |
H A D | marvell_pem_pmu.c | 238 u64 prev_count, new_count; in pem_perf_event_update() local 242 new_count = pem_perf_read_counter(pmu, event, hwc->idx); in pem_perf_event_update() 243 } while (local64_xchg(&hwc->prev_count, new_count) != prev_count); in pem_perf_event_update() 245 local64_add((new_count - prev_count), &event->count); in pem_perf_event_update()
|
H A D | marvell_cn10k_ddr_pmu.c | 595 u64 prev_count, new_count, mask; in cn10k_ddr_perf_event_update() local 599 new_count = cn10k_ddr_perf_read_counter(pmu, hwc->idx); in cn10k_ddr_perf_event_update() 600 } while (local64_xchg(&hwc->prev_count, new_count) != prev_count); in cn10k_ddr_perf_event_update() 604 local64_add((new_count - prev_count) & mask, &event->count); in cn10k_ddr_perf_event_update() 856 u64 prev_count, new_count; in cn10k_ddr_pmu_overflow_handler() local 864 new_count = cn10k_ddr_perf_read_counter(pmu, hwc->idx); in cn10k_ddr_pmu_overflow_handler() 869 if (new_count < prev_count) in cn10k_ddr_pmu_overflow_handler() 877 new_count = cn10k_ddr_perf_read_counter(pmu, hwc->idx); in cn10k_ddr_pmu_overflow_handler() 882 if (new_count < prev_count) in cn10k_ddr_pmu_overflow_handler()
|
H A D | arm_dsu_pmu.c | 334 u64 delta, prev_count, new_count; in dsu_pmu_event_update() local 339 new_count = dsu_pmu_read_counter(event); in dsu_pmu_event_update() 340 } while (local64_cmpxchg(&hwc->prev_count, prev_count, new_count) != in dsu_pmu_event_update() 342 delta = (new_count - prev_count) & DSU_PMU_COUNTER_MASK(hwc->idx); in dsu_pmu_event_update()
|
H A D | arm_dmc620_pmu.c | 327 u64 delta, prev_count, new_count; in dmc620_pmu_event_update() local 332 new_count = dmc620_pmu_read_counter(event); in dmc620_pmu_event_update() 334 prev_count, new_count) != prev_count); in dmc620_pmu_event_update() 335 delta = (new_count - prev_count) & DMC620_CNT_MAX_PERIOD; in dmc620_pmu_event_update()
|
/linux/kernel/trace/ |
H A D | trace_functions.c | 645 long new_count; in ftrace_stacktrace_count() local 668 new_count = old_count - 1; in ftrace_stacktrace_count() 669 new_count = cmpxchg(count, old_count, new_count); in ftrace_stacktrace_count() 670 if (new_count == old_count) in ftrace_stacktrace_count() 676 } while (new_count != old_count); in ftrace_stacktrace_count()
|
/linux/include/linux/ |
H A D | parman.h | 58 int (*resize)(void *priv, unsigned long new_count);
|
/linux/drivers/vfio/pci/hisilicon/ |
H A D | hisi_acc_vfio_pci.c | 1255 size_t *new_count) in hisi_acc_pci_rw_access_check() argument 1269 *new_count = min(count, (size_t)(end - pos)); in hisi_acc_pci_rw_access_check() 1303 size_t new_count = count; in hisi_acc_vfio_pci_write() local 1306 ret = hisi_acc_pci_rw_access_check(core_vdev, count, ppos, &new_count); in hisi_acc_vfio_pci_write() 1310 return vfio_pci_core_write(core_vdev, buf, new_count, ppos); in hisi_acc_vfio_pci_write() 1317 size_t new_count = count; in hisi_acc_vfio_pci_read() local 1320 ret = hisi_acc_pci_rw_access_check(core_vdev, count, ppos, &new_count); in hisi_acc_vfio_pci_read() 1324 return vfio_pci_core_read(core_vdev, buf, new_count, ppos); in hisi_acc_vfio_pci_read()
|
/linux/drivers/iommu/intel/ |
H A D | perfmon.c | 305 u64 prev_count, new_count, delta; in iommu_pmu_event_update() local 310 new_count = dmar_readq(iommu_event_base(iommu_pmu, hwc->idx)); in iommu_pmu_event_update() 311 if (local64_xchg(&hwc->prev_count, new_count) != prev_count) in iommu_pmu_event_update() 318 delta = (new_count << shift) - (prev_count << shift); in iommu_pmu_event_update()
|
/linux/drivers/net/ethernet/microsoft/mana/ |
H A D | mana_ethtool.c | 395 unsigned int new_count = channels->combined_count; in mana_set_channels() local 399 err = mana_pre_alloc_rxbufs(apc, ndev->mtu, new_count); in mana_set_channels() 411 apc->num_queues = new_count; in mana_set_channels()
|
/linux/drivers/md/ |
H A D | dm-thin-metadata.c | 2028 static int __resize_space_map(struct dm_space_map *sm, dm_block_t new_count) in __resize_space_map() argument 2037 if (new_count == old_count) in __resize_space_map() 2040 if (new_count < old_count) { in __resize_space_map() 2045 return dm_sm_extend(sm, new_count - old_count); in __resize_space_map() 2048 int dm_pool_resize_data_dev(struct dm_pool_metadata *pmd, dm_block_t new_count) in dm_pool_resize_data_dev() argument 2054 r = __resize_space_map(pmd->data_sm, new_count); in dm_pool_resize_data_dev() 2060 int dm_pool_resize_metadata_dev(struct dm_pool_metadata *pmd, dm_block_t new_count) in dm_pool_resize_metadata_dev() argument 2066 r = __resize_space_map(pmd->metadata_sm, new_count); in dm_pool_resize_metadata_dev()
|
/linux/drivers/scsi/isci/ |
H A D | remote_node_context.c | 357 u32 new_count = rnc->suspend_count + 1; in sci_remote_node_context_tx_rx_suspended_state_enter() local 359 if (new_count == 0) in sci_remote_node_context_tx_rx_suspended_state_enter() 362 rnc->suspend_count = new_count; in sci_remote_node_context_tx_rx_suspended_state_enter()
|
/linux/drivers/char/ |
H A D | random.c | 1120 unsigned int new_count; in add_interrupt_randomness() local 1124 new_count = ++fast_pool->count; in add_interrupt_randomness() 1126 if (new_count & MIX_INFLIGHT) in add_interrupt_randomness() 1129 if (new_count < 1024 && !time_is_before_jiffies(fast_pool->last + HZ)) in add_interrupt_randomness()
|
/linux/arch/sparc/kernel/ |
H A D | ds.c | 896 u64 new_count; in register_services() local 901 new_count = sched_clock() & 0xffffffff; in register_services() 902 cp->handle = ((u64) i << 32) | new_count; in register_services()
|