| /drivers/remoteproc/ |
| A D | ti_k3_r5_remoteproc.c | 307 struct k3_r5_cluster *cluster = core->cluster; in k3_r5_rproc_prepare() local 405 struct k3_r5_cluster *cluster = core->cluster; in k3_r5_rproc_unprepare() local 468 struct k3_r5_cluster *cluster = core->cluster; in k3_r5_rproc_start() local 533 struct k3_r5_cluster *cluster = core->cluster; in k3_r5_rproc_stop() local 645 struct k3_r5_cluster *cluster = core->cluster; in k3_r5_rproc_configure() local 780 struct k3_r5_cluster *cluster = core->cluster; in k3_r5_adjust_tcm_sizes() local 818 struct k3_r5_cluster *cluster = core->cluster; in k3_r5_rproc_configure_mode() local 1326 core->cluster = cluster; in k3_r5_cluster_of_init() 1353 cluster = devm_kzalloc(dev, sizeof(*cluster), GFP_KERNEL); in k3_r5_probe() 1354 if (!cluster) in k3_r5_probe() [all …]
|
| A D | mtk_scp.c | 70 struct mtk_scp_of_cluster *scp_cluster = scp->cluster; in scp_wdt_handler() 175 val = readl(scp->cluster->reg_base + MT8183_SW_RSTN); in mt8183_scp_reset_assert() 177 writel(val, scp->cluster->reg_base + MT8183_SW_RSTN); in mt8183_scp_reset_assert() 184 val = readl(scp->cluster->reg_base + MT8183_SW_RSTN); in mt8183_scp_reset_deassert() 186 writel(val, scp->cluster->reg_base + MT8183_SW_RSTN); in mt8183_scp_reset_deassert() 221 scp->cluster->reg_base + MT8183_SCP_TO_HOST); in mt8183_scp_irq_handler() 779 if (scp->cluster->l1tcm_size && in mt8192_scp_da_to_va() 780 da >= scp->cluster->l1tcm_phys && in mt8192_scp_da_to_va() 781 (da + len) <= scp->cluster->l1tcm_phys + scp->cluster->l1tcm_size) { in mt8192_scp_da_to_va() 782 offset = da - scp->cluster->l1tcm_phys; in mt8192_scp_da_to_va() [all …]
|
| A D | xlnx_r5_remoteproc.c | 1409 cluster->mode = cluster_mode; in zynqmp_r5_cluster_init() 1416 cluster->core_count = 0; in zynqmp_r5_cluster_init() 1417 cluster->r5_cores = NULL; in zynqmp_r5_cluster_init() 1454 if (!cluster) in zynqmp_r5_cluster_exit() 1467 kfree(cluster->r5_cores); in zynqmp_r5_cluster_exit() 1468 kfree(cluster); in zynqmp_r5_cluster_exit() 1526 cluster = kzalloc(sizeof(*cluster), GFP_KERNEL); in zynqmp_r5_remoteproc_probe() 1527 if (!cluster) in zynqmp_r5_remoteproc_probe() 1530 cluster->dev = dev; in zynqmp_r5_remoteproc_probe() 1535 kfree(cluster); in zynqmp_r5_remoteproc_probe() [all …]
|
| A D | mtk_scp_ipi.c | 183 ret = readl_poll_timeout_atomic(scp->cluster->reg_base + scp->data->host_to_scp_reg, in scp_ipi_send() 198 scp->cluster->reg_base + scp->data->host_to_scp_reg); in scp_ipi_send()
|
| A D | mtk_common.h | 157 struct mtk_scp_of_cluster *cluster; member
|
| /drivers/perf/ |
| A D | qcom_l2_pmu.c | 489 if (!cluster) { in l2_cache_event_init() 761 return cluster; in l2_cache_associate_cpu_with_cluster() 774 if (!cluster) { in l2cache_pmu_online_cpu() 777 if (!cluster) { in l2cache_pmu_online_cpu() 792 cluster->on_cpu = cpu; in l2cache_pmu_online_cpu() 810 if (!cluster) in l2cache_pmu_offline_cpu() 819 cluster->on_cpu = -1; in l2cache_pmu_offline_cpu() 853 cluster = devm_kzalloc(&pdev->dev, sizeof(*cluster), GFP_KERNEL); in l2_cache_pmu_probe_cluster() 854 if (!cluster) in l2_cache_pmu_probe_cluster() 863 cluster->irq = irq; in l2_cache_pmu_probe_cluster() [all …]
|
| /drivers/cpufreq/ |
| A D | vexpress-spc-cpufreq.c | 45 #define ACTUAL_FREQ(cluster, freq) ((cluster == A7_CLUSTER) ? freq << 1 : freq) argument 46 #define VIRT_FREQ(cluster, freq) ((cluster == A7_CLUSTER) ? freq >> 1 : freq) argument 284 if (!freq_table[cluster]) in _put_cluster_clk_and_freq_table() 287 clk_put(clk[cluster]); in _put_cluster_clk_and_freq_table() 300 if (cluster < MAX_CLUSTERS) in put_cluster_clk_and_freq_table() 313 kfree(freq_table[cluster]); in put_cluster_clk_and_freq_table() 322 if (freq_table[cluster]) in _get_cluster_clk_and_freq_table() 338 if (!IS_ERR(clk[cluster])) in _get_cluster_clk_and_freq_table() 343 ret = PTR_ERR(clk[cluster]); in _get_cluster_clk_and_freq_table() 348 cluster); in _get_cluster_clk_and_freq_table() [all …]
|
| A D | tegra186-cpufreq.c | 78 policy->freq_table = data->clusters[cluster].table; in tegra186_cpufreq_init() 84 if (data->cpus[cpu].bpmp_cluster_id == cluster) in tegra186_cpufreq_init() 107 struct tegra186_cpufreq_cluster *cluster; in tegra186_cpufreq_get() local 119 cluster = &data->clusters[cluster_id]; in tegra186_cpufreq_get() 122 return (cluster->ref_clk_khz * ndiv) / cluster->div; in tegra186_cpufreq_get() 193 cluster->ref_clk_khz = data->ref_clk_hz / 1000; in init_vhint_table() 194 cluster->div = data->pdiv * data->mdiv; in init_vhint_table() 213 point->frequency = (cluster->ref_clk_khz * ndiv) / cluster->div; in init_vhint_table() 251 cluster->table = init_vhint_table(pdev, bpmp, cluster, i); in tegra186_cpufreq_probe() 252 if (IS_ERR(cluster->table)) { in tegra186_cpufreq_probe() [all …]
|
| /drivers/clk/mvebu/ |
| A D | ap-cpu-clk.c | 135 unsigned int cluster; member 151 (clk->cluster * clk->pll_regs->cluster_offset); in ap_cpu_clk_recalc_rate() 167 (clk->cluster * clk->pll_regs->cluster_offset); in ap_cpu_clk_set_rate() 169 (clk->cluster * clk->pll_regs->cluster_offset); in ap_cpu_clk_set_rate() 171 (clk->cluster * clk->pll_regs->cluster_offset); in ap_cpu_clk_set_rate() 198 clk->cluster * in ap_cpu_clk_set_rate() 315 ap_cpu_clk[cluster_index].cluster = cluster_index; in ap_cpu_clock_probe()
|
| /drivers/gpu/drm/msm/adreno/ |
| A D | a6xx_gpu_state.c | 693 const struct a6xx_cluster *cluster, in a6xx_get_cluster() argument 702 u32 id = cluster->id; in a6xx_get_cluster() 714 if (cluster->sel_reg) in a6xx_get_cluster() 715 in += CRASHDUMP_WRITE(in, cluster->sel_reg, cluster->sel_val); in a6xx_get_cluster() 746 obj->handle = cluster; in a6xx_get_cluster() 763 if (cluster->sel) in a7xx_get_cluster() 764 in += CRASHDUMP_WRITE(in, cluster->sel->cd_reg, cluster->sel->val); in a7xx_get_cluster() 791 obj->handle = cluster; in a7xx_get_cluster() 1860 if (cluster) { in a6xx_show_cluster() 1862 a6xx_show_cluster_data(cluster->registers, cluster->count, in a6xx_show_cluster() [all …]
|
| /drivers/media/v4l2-core/ |
| A D | v4l2-ctrls-request.c | 363 struct v4l2_ctrl *master = ctrl->cluster[0]; in v4l2_ctrl_request_complete() 370 cur_to_new(master->cluster[i]); in v4l2_ctrl_request_complete() 428 struct v4l2_ctrl *master = ctrl->cluster[0]; in v4l2_ctrl_request_setup() 441 if (master->cluster[i]) { in v4l2_ctrl_request_setup() 443 find_ref(hdl, master->cluster[i]->id); in v4l2_ctrl_request_setup() 457 if (master->cluster[i]) { in v4l2_ctrl_request_setup() 459 find_ref(hdl, master->cluster[i]->id); in v4l2_ctrl_request_setup() 466 master->cluster[i]->is_new = 1; in v4l2_ctrl_request_setup()
|
| A D | v4l2-ctrls-api.c | 287 if (ctrl->cluster[0]->ncontrols > 1) in prepare_ext_ctrls() 289 if (ctrl->cluster[0] != ctrl) in prepare_ext_ctrls() 473 cur_to_new(master->cluster[j]); in v4l2_g_ext_ctrls_common() 643 if (master->cluster[j]) in try_set_ext_ctrls_common() 644 master->cluster[j]->is_new = 0; in try_set_ext_ctrls_common() 701 find_ref(hdl, master->cluster[j]->id); in try_set_ext_ctrls_common() 770 struct v4l2_ctrl *master = ctrl->cluster[0]; in get_ctrl() 788 cur_to_new(master->cluster[i]); in get_ctrl() 819 struct v4l2_ctrl *master = ctrl->cluster[0]; in set_ctrl() 825 if (master->cluster[i]) in set_ctrl() [all …]
|
| A D | v4l2-ctrls-core.c | 1417 if (!is_cur_manual(ctrl->cluster[0])) { in new_to_cur() 1419 if (ctrl->cluster[0]->has_volatiles) in new_to_cur() 1826 ctrl->cluster = &new_ref->ctrl; in handler_new_ref() 2377 controls[i]->cluster = controls; in v4l2_ctrl_cluster() 2422 cur_to_new(master->cluster[i]); in update_from_auto_cluster() 2425 if (master->cluster[i]) in update_from_auto_cluster() 2426 master->cluster[i]->is_new = 1; in update_from_auto_cluster() 2602 if (master->cluster[i]) { in __v4l2_ctrl_handler_setup() 2603 cur_to_new(master->cluster[i]); in __v4l2_ctrl_handler_setup() 2604 master->cluster[i]->is_new = 1; in __v4l2_ctrl_handler_setup() [all …]
|
| /drivers/media/platform/microchip/ |
| A D | microchip-isc-base.c | 1480 if (ctrl->cluster[ISC_CTRL_R_GAIN]->is_new) in isc_s_awb_ctrl() 1489 if (ctrl->cluster[ISC_CTRL_R_OFF]->is_new) in isc_s_awb_ctrl() 1491 if (ctrl->cluster[ISC_CTRL_B_OFF]->is_new) in isc_s_awb_ctrl() 1549 ctrl->cluster[ISC_CTRL_R_GAIN]->val = in isc_g_volatile_awb_ctrl() 1551 ctrl->cluster[ISC_CTRL_B_GAIN]->val = in isc_g_volatile_awb_ctrl() 1553 ctrl->cluster[ISC_CTRL_GR_GAIN]->val = in isc_g_volatile_awb_ctrl() 1555 ctrl->cluster[ISC_CTRL_GB_GAIN]->val = in isc_g_volatile_awb_ctrl() 1558 ctrl->cluster[ISC_CTRL_R_OFF]->val = in isc_g_volatile_awb_ctrl() 1560 ctrl->cluster[ISC_CTRL_B_OFF]->val = in isc_g_volatile_awb_ctrl() 1562 ctrl->cluster[ISC_CTRL_GR_OFF]->val = in isc_g_volatile_awb_ctrl() [all …]
|
| /drivers/staging/media/deprecated/atmel/ |
| A D | atmel-isc-base.c | 1532 if (ctrl->cluster[ISC_CTRL_R_GAIN]->is_new) in isc_s_awb_ctrl() 1541 if (ctrl->cluster[ISC_CTRL_R_OFF]->is_new) in isc_s_awb_ctrl() 1543 if (ctrl->cluster[ISC_CTRL_B_OFF]->is_new) in isc_s_awb_ctrl() 1602 ctrl->cluster[ISC_CTRL_R_GAIN]->val = in isc_g_volatile_awb_ctrl() 1604 ctrl->cluster[ISC_CTRL_B_GAIN]->val = in isc_g_volatile_awb_ctrl() 1606 ctrl->cluster[ISC_CTRL_GR_GAIN]->val = in isc_g_volatile_awb_ctrl() 1608 ctrl->cluster[ISC_CTRL_GB_GAIN]->val = in isc_g_volatile_awb_ctrl() 1611 ctrl->cluster[ISC_CTRL_R_OFF]->val = in isc_g_volatile_awb_ctrl() 1613 ctrl->cluster[ISC_CTRL_B_OFF]->val = in isc_g_volatile_awb_ctrl() 1615 ctrl->cluster[ISC_CTRL_GR_OFF]->val = in isc_g_volatile_awb_ctrl() [all …]
|
| /drivers/block/drbd/ |
| A D | Kconfig | 33 shared-nothing cluster. Needless to say, on top of dual-Primary 34 DRBD utilizing a cluster file system is necessary to maintain for 37 For automatic failover you need a cluster manager (e.g. heartbeat).
|
| /drivers/crypto/intel/qat/qat_common/ |
| A D | adf_rl.c | 165 if (rl_data->cluster[i] && rl_data->cluster[i]->parent == root) in find_parent() 166 return rl_data->cluster[i]; in find_parent() 188 *sla_arr = rl_data->cluster; in adf_rl_get_sla_arr_of_type() 505 sla_list = rl_data->cluster; in get_next_free_node_id() 969 if (!rl_data->cluster[i]) in adf_rl_get_capability_remaining() 972 if (rl_data->cluster[i]->srv == srv) { in adf_rl_get_capability_remaining() 973 sla = rl_data->cluster[i]; in adf_rl_get_capability_remaining()
|
| A D | adf_rl.h | 115 struct rl_sla *cluster[RL_CLUSTER_MAX]; member
|
| /drivers/cpuidle/ |
| A D | cpuidle-big_little.c | 106 unsigned int cluster = MPIDR_AFFINITY_LEVEL(mpidr, 1); in bl_powerdown_finisher() local 109 mcpm_set_entry_vector(cpu, cluster, cpu_resume); in bl_powerdown_finisher()
|
| /drivers/base/ |
| A D | arch_topology.c | 544 static int __init parse_cluster(struct device_node *cluster, int package_id, in parse_cluster() argument 562 of_get_child_by_name(cluster, name); in parse_cluster() 581 of_get_child_by_name(cluster, name); in parse_cluster() 599 cluster, name); in parse_cluster() 607 pr_warn("%pOF: empty cluster\n", cluster); in parse_cluster()
|
| /drivers/mailbox/ |
| A D | mailbox-mchp-ipc-sbi.c | 81 u8 cluster; member 187 status_msg.cluster = hartid; in mchp_ipc_cluster_aggr_isr() 214 if (chan_index >= status_msg.cluster) in mchp_ipc_cluster_aggr_isr() 222 chan_id = status_msg.cluster * (NUM_CHANS_PER_CLUSTER + chan_index); in mchp_ipc_cluster_aggr_isr()
|
| /drivers/md/ |
| A D | dm-verity-target.c | 654 unsigned int cluster = READ_ONCE(dm_verity_prefetch_cluster); in verity_prefetch_io() local 656 cluster >>= v->data_dev_block_bits; in verity_prefetch_io() 657 if (unlikely(!cluster)) in verity_prefetch_io() 660 if (unlikely(cluster & (cluster - 1))) in verity_prefetch_io() 661 cluster = 1 << __fls(cluster); in verity_prefetch_io() 663 hash_block_start &= ~(sector_t)(cluster - 1); in verity_prefetch_io() 664 hash_block_end |= cluster - 1; in verity_prefetch_io()
|
| /drivers/target/tcm_remote/ |
| A D | Kconfig | 8 of TPG/ACL/LUN on peer nodes in a cluster.
|
| /drivers/pinctrl/nomadik/ |
| A D | pinctrl-abx500.c | 225 struct abx500_gpio_irq_cluster *cluster = in abx500_gpio_to_irq() local 228 if (gpio >= cluster->start && gpio <= cluster->end) { in abx500_gpio_to_irq() 235 hwirq = gpio - cluster->start + cluster->to_irq; in abx500_gpio_to_irq()
|
| /drivers/net/ethernet/packetengines/ |
| A D | Kconfig | 37 used by the Beowulf Linux cluster project. See
|