| /drivers/net/wireless/intel/iwlwifi/dvm/ |
| A D | main.c | 609 priv->contexts[i].ctxid = i; in iwl_init_context() 611 priv->contexts[IWL_RXON_CTX_BSS].always_active = true; in iwl_init_context() 612 priv->contexts[IWL_RXON_CTX_BSS].is_active = true; in iwl_init_context() 622 priv->contexts[IWL_RXON_CTX_BSS].interface_modes = in iwl_init_context() 628 memcpy(priv->contexts[IWL_RXON_CTX_BSS].ac_to_queue, in iwl_init_context() 630 memcpy(priv->contexts[IWL_RXON_CTX_BSS].ac_to_fifo, in iwl_init_context() 634 priv->contexts[IWL_RXON_CTX_PAN].rxon_timing_cmd = in iwl_init_context() 636 priv->contexts[IWL_RXON_CTX_PAN].rxon_assoc_cmd = in iwl_init_context() 643 priv->contexts[IWL_RXON_CTX_PAN].interface_modes = in iwl_init_context() 649 memcpy(priv->contexts[IWL_RXON_CTX_PAN].ac_to_queue, in iwl_init_context() [all …]
|
| A D | rxon.c | 301 priv->contexts[IWL_RXON_CTX_BSS].vif && in iwl_send_rxon_timing() 302 priv->contexts[IWL_RXON_CTX_BSS].vif->bss_conf.beacon_int) { in iwl_send_rxon_timing() 304 priv->contexts[IWL_RXON_CTX_BSS].timing.beacon_interval; in iwl_send_rxon_timing() 308 priv->contexts[IWL_RXON_CTX_PAN].vif && in iwl_send_rxon_timing() 309 priv->contexts[IWL_RXON_CTX_PAN].vif->bss_conf.beacon_int && in iwl_send_rxon_timing() 313 priv->contexts[IWL_RXON_CTX_PAN].timing.beacon_interval; in iwl_send_rxon_timing() 387 struct iwl_rxon_context *ctx = &priv->contexts[IWL_RXON_CTX_BSS]; in iwl_set_tx_power() 523 ctx_bss = &priv->contexts[IWL_RXON_CTX_BSS]; in iwlagn_set_pan_params() 524 ctx_pan = &priv->contexts[IWL_RXON_CTX_PAN]; in iwlagn_set_pan_params() 891 struct iwl_rxon_context *ctx = &priv->contexts[ctxid]; in iwl_print_rx_config_cmd()
|
| A D | dev.h | 721 struct iwl_rxon_context contexts[NUM_IWL_RXON_CTX]; member 903 for (ctx = &priv->contexts[IWL_RXON_CTX_BSS]; \ 904 ctx < &priv->contexts[NUM_IWL_RXON_CTX]; ctx++) \ 915 return iwl_is_associated_ctx(&priv->contexts[ctxid]); in iwl_is_associated()
|
| A D | mac80211.c | 337 if (priv->contexts[IWL_RXON_CTX_BSS].vif != vif) in iwlagn_mac_set_rekey_data() 357 struct iwl_rxon_context *ctx = &priv->contexts[IWL_RXON_CTX_BSS]; in iwlagn_mac_suspend() 421 struct iwl_rxon_context *ctx = &priv->contexts[IWL_RXON_CTX_BSS]; in iwlagn_mac_resume() 520 if (resume_data.valid && priv->contexts[IWL_RXON_CTX_BSS].vif) { in iwlagn_mac_resume() 940 struct iwl_rxon_context *ctx = &priv->contexts[IWL_RXON_CTX_BSS]; in iwlagn_mac_channel_switch() 1014 struct iwl_rxon_context *ctx = &priv->contexts[IWL_RXON_CTX_BSS]; in iwl_chswitch_done()
|
| A D | devices.c | 394 struct iwl_rxon_context *ctx = &priv->contexts[IWL_RXON_CTX_BSS]; in iwl5000_hw_channel_switch() 555 struct iwl_rxon_context *ctx = &priv->contexts[IWL_RXON_CTX_BSS]; in iwl6000_hw_channel_switch()
|
| /drivers/irqchip/ |
| A D | irq-aclint-sswi.c | 64 u32 contexts = of_irq_count(to_of_node(fwnode)); in aclint_sswi_parse_irq() local 66 if (!(contexts)) { in aclint_sswi_parse_irq() 71 for (u32 i = 0; i < contexts; i++) { in aclint_sswi_parse_irq() 98 pr_info("%pfwP: register %u CPU%s\n", fwnode, contexts, str_plural(contexts)); in aclint_sswi_parse_irq()
|
| /drivers/gpu/drm/i915/ |
| A D | i915_sysfs.c | 79 spin_lock(&i915->gem.contexts.lock); in i915_l3_read() 84 spin_unlock(&i915->gem.contexts.lock); in i915_l3_read() 112 spin_lock(&i915->gem.contexts.lock); in i915_l3_write() 125 list_for_each_entry(ctx, &i915->gem.contexts.list, link) in i915_l3_write() 128 spin_unlock(&i915->gem.contexts.lock); in i915_l3_write()
|
| /drivers/gpu/drm/tegra/ |
| A D | uapi.c | 55 xa_for_each(&file->contexts, id, context) in tegra_drm_uapi_close_file() 61 xa_destroy(&file->contexts); in tegra_drm_uapi_close_file() 135 err = xa_alloc(&fpriv->contexts, &args->context, context, XA_LIMIT(1, U32_MAX), in tegra_drm_ioctl_channel_open() 170 context = xa_load(&fpriv->contexts, args->context); in tegra_drm_ioctl_channel_close() 176 xa_erase(&fpriv->contexts, args->context); in tegra_drm_ioctl_channel_close() 200 context = xa_load(&fpriv->contexts, args->context); in tegra_drm_ioctl_channel_map() 281 context = xa_load(&fpriv->contexts, args->context); in tegra_drm_ioctl_channel_unmap()
|
| A D | uapi.h | 23 struct xarray contexts; member
|
| /drivers/gpu/drm/i915/pxp/ |
| A D | intel_pxp.c | 498 spin_lock_irq(&i915->gem.contexts.lock); in intel_pxp_invalidate() 499 list_for_each_entry_safe(ctx, cn, &i915->gem.contexts.list, link) { in intel_pxp_invalidate() 511 spin_unlock_irq(&i915->gem.contexts.lock); in intel_pxp_invalidate() 537 spin_lock_irq(&i915->gem.contexts.lock); in intel_pxp_invalidate() 541 spin_unlock_irq(&i915->gem.contexts.lock); in intel_pxp_invalidate()
|
| /drivers/staging/media/ipu7/ |
| A D | ipu7-boot.c | 39 static const struct ipu7_boot_context contexts[IPU_SUBSYS_NUM] = { variable 86 const struct ipu7_boot_context *ctx = &contexts[adev->subsys]; in ipu7_boot_cell_reset() 136 const struct ipu7_boot_context *ctx = &contexts[adev->subsys]; in ipu7_boot_cell_start() 154 const struct ipu7_boot_context *ctx = &contexts[adev->subsys]; in ipu7_boot_cell_stop() 179 const struct ipu7_boot_context *ctx = &contexts[adev->subsys]; in ipu7_boot_cell_init()
|
| /drivers/media/platform/ti/omap3isp/ |
| A D | ispcsi2.c | 220 struct isp_csi2_ctx_cfg *ctx = &csi2->contexts[0]; in csi2_set_outaddr() 250 struct isp_csi2_ctx_cfg *ctx = &csi2->contexts[ctxnum]; in csi2_ctx_enable() 561 if (csi2->contexts[0].enabled || csi2->ctrl.if_enable) in csi2_configure() 599 csi2->contexts[0].format_id = csi2_ctx_map_format(csi2); in csi2_configure() 602 csi2->contexts[0].data_offset = 0; in csi2_configure() 604 csi2->contexts[0].data_offset = csi2->video_out.bpl_value; in csi2_configure() 613 csi2->contexts[0].eof_enabled = 1; in csi2_configure() 614 csi2->contexts[0].eol_enabled = 1; in csi2_configure() 623 csi2_ctx_config(isp, csi2, &csi2->contexts[0]); in csi2_configure() 785 csi2_isr_ctx(csi2, &csi2->contexts[0]); in omap3isp_csi2_isr()
|
| A D | ispcsi2.h | 137 struct isp_csi2_ctx_cfg contexts[ISP_CSI2_MAX_CTX_NUM + 1]; member
|
| /drivers/gpu/drm/imagination/ |
| A D | pvr_context.c | 350 list_add_tail(&ctx->file_link, &pvr_file->contexts); in pvr_context_create() 449 ctx = list_first_entry(&pvr_file->contexts, struct pvr_context, file_link); in pvr_destroy_contexts_for_file() 451 while (!list_entry_is_head(ctx, &pvr_file->contexts, file_link)) { in pvr_destroy_contexts_for_file() 462 ctx = list_first_entry(&pvr_file->contexts, struct pvr_context, file_link); in pvr_destroy_contexts_for_file()
|
| /drivers/gpu/drm/i915/selftests/ |
| A D | i915_vma.c | 109 struct list_head *contexts) in create_vmas() argument 117 list_for_each_entry(ctx, contexts, link) { in create_vmas() 158 LIST_HEAD(contexts); in igt_vma_create() 184 list_move(&ctx->link, &contexts); in igt_vma_create() 187 err = create_vmas(i915, &objects, &contexts); in igt_vma_create() 197 list_for_each_entry_safe(ctx, cn, &contexts, link) { in igt_vma_create() 207 err = create_vmas(i915, &objects, &contexts); in igt_vma_create() 209 list_for_each_entry_safe(ctx, cn, &contexts, link) { in igt_vma_create()
|
| A D | i915_request.c | 284 struct i915_gem_context **contexts; member 478 t.contexts = kcalloc(t.ncontexts, sizeof(*t.contexts), GFP_KERNEL); in mock_breadcrumbs_smoketest() 479 if (!t.contexts) { in mock_breadcrumbs_smoketest() 486 if (!t.contexts[n]) { in mock_breadcrumbs_smoketest() 532 if (!t.contexts[n]) in mock_breadcrumbs_smoketest() 534 mock_context_close(t.contexts[n]); in mock_breadcrumbs_smoketest() 536 kfree(t.contexts); in mock_breadcrumbs_smoketest() 1771 sizeof(*smoke[0].contexts), in live_breadcrumbs_smoketest() 1773 if (!smoke[0].contexts) { in live_breadcrumbs_smoketest() 1780 if (IS_ERR(smoke[0].contexts[n])) { in live_breadcrumbs_smoketest() [all …]
|
| /drivers/gpu/drm/xe/ |
| A D | Kconfig.profile | 19 contexts by guc scheduling. 25 contexts by guc scheduling.
|
| /drivers/md/dm-vdo/ |
| A D | dedupe.h | 67 struct dedupe_context contexts[MAXIMUM_VDO_USER_VIOS]; member
|
| A D | vdo.h | 292 unsigned int queue_count, void *contexts[]);
|
| /drivers/gpu/drm/i915/gem/selftests/ |
| A D | mock_context.c | 74 init_contexts(&i915->gem.contexts); in mock_init_contexts()
|
| /drivers/net/ethernet/marvell/octeontx2/ |
| A D | Kconfig | 29 NPA Aura/Pool contexts.
|
| /drivers/net/wireless/intel/iwlwifi/mvm/ |
| A D | ops.c | 1686 u8 contexts) in iwl_mvm_async_handlers_by_context() argument 1698 if (!(BIT(entry->context) & contexts)) in iwl_mvm_async_handlers_by_context() 1722 u8 contexts = BIT(RX_HANDLER_ASYNC_LOCKED_WIPHY); in iwl_mvm_async_handlers_wiphy_wk() local 1724 iwl_mvm_async_handlers_by_context(mvm, contexts); in iwl_mvm_async_handlers_wiphy_wk() 1731 u8 contexts = BIT(RX_HANDLER_ASYNC_LOCKED) | in iwl_mvm_async_handlers_wk() local 1734 iwl_mvm_async_handlers_by_context(mvm, contexts); in iwl_mvm_async_handlers_wk()
|
| /drivers/misc/ocxl/ |
| A D | sysfs.c | 92 __ATTR_RO(contexts),
|
| /drivers/gpu/drm/i915/gem/ |
| A D | i915_gem_context.c | 1290 spin_lock(&ctx->i915->gem.contexts.lock); in i915_gem_context_release_work() 1292 spin_unlock(&ctx->i915->gem.contexts.lock); in i915_gem_context_release_work() 1708 init_contexts(&i915->gem.contexts); in i915_gem_init__contexts() 1734 spin_lock(&i915->gem.contexts.lock); in gem_context_register() 1735 list_add_tail(&ctx->link, &i915->gem.contexts.list); in gem_context_register() 1736 spin_unlock(&i915->gem.contexts.lock); in gem_context_register()
|
| /drivers/gpu/drm/i915/gt/ |
| A D | selftest_execlists.c | 3471 struct i915_gem_context **contexts; member 3482 return smoke->contexts[i915_prandom_u32_max_state(smoke->ncontext, in smoke_context() 3648 smoke.contexts = kmalloc_array(smoke.ncontext, in live_preempt_smoke() 3649 sizeof(*smoke.contexts), in live_preempt_smoke() 3651 if (!smoke.contexts) in live_preempt_smoke() 3678 smoke.contexts[n] = kernel_context(smoke.gt->i915, NULL); in live_preempt_smoke() 3679 if (!smoke.contexts[n]) in live_preempt_smoke() 3698 if (!smoke.contexts[n]) in live_preempt_smoke() 3700 kernel_context_close(smoke.contexts[n]); in live_preempt_smoke() 3706 kfree(smoke.contexts); in live_preempt_smoke()
|