| /drivers/firmware/broadcom/ |
| A D | bcm47xx_sprom.c | 75 fallback); \ 118 bool fallback) in nvram_read_leddc() argument 192 bool fb = fallback; in bcm47xx_sprom_fill_auto() 513 bool fb = fallback; in bcm47xx_fill_sprom_ethernet() 517 fallback); in bcm47xx_fill_sprom_ethernet() 519 fallback); in bcm47xx_fill_sprom_ethernet() 523 fallback); in bcm47xx_fill_sprom_ethernet() 525 fallback); in bcm47xx_fill_sprom_ethernet() 556 bool fallback) in bcm47xx_fill_board_data() argument 565 bool fallback) in bcm47xx_fill_sprom() argument [all …]
|
| /drivers/crypto/ |
| A D | geode-aes.c | 130 tctx->fallback.cip->base.crt_flags |= in geode_setkey_cip() 155 crypto_skcipher_clear_flags(tctx->fallback.skcipher, in geode_setkey_skcipher() 157 crypto_skcipher_set_flags(tctx->fallback.skcipher, in geode_setkey_skcipher() 197 tctx->fallback.cip = crypto_alloc_cipher(name, 0, in fallback_init_cip() 200 if (IS_ERR(tctx->fallback.cip)) { in fallback_init_cip() 202 return PTR_ERR(tctx->fallback.cip); in fallback_init_cip() 212 crypto_free_cipher(tctx->fallback.cip); in fallback_exit_cip() 243 tctx->fallback.skcipher = in geode_init_skcipher() 246 if (IS_ERR(tctx->fallback.skcipher)) { in geode_init_skcipher() 248 return PTR_ERR(tctx->fallback.skcipher); in geode_init_skcipher() [all …]
|
| A D | atmel-ecc.c | 42 struct crypto_kpp *fallback; member 145 kpp_request_set_tfm(req, ctx->fallback); in atmel_ecdh_generate_public_key() 174 kpp_request_set_tfm(req, ctx->fallback); in atmel_ecdh_compute_shared_secret() 250 struct crypto_kpp *fallback; in atmel_ecdh_init_tfm() local 261 if (IS_ERR(fallback)) { in atmel_ecdh_init_tfm() 263 alg, PTR_ERR(fallback)); in atmel_ecdh_init_tfm() 264 return PTR_ERR(fallback); in atmel_ecdh_init_tfm() 268 ctx->fallback = fallback; in atmel_ecdh_init_tfm() 278 crypto_free_kpp(ctx->fallback); in atmel_ecdh_exit_tfm() 286 if (ctx->fallback) in atmel_ecdh_max_size() [all …]
|
| A D | padlock-sha.c | 24 struct crypto_ahash *fallback; member 61 HASH_REQUEST_ON_STACK(req, ctx->fallback); in padlock_sha_update() 106 HASH_REQUEST_ON_STACK(req, ctx->fallback); in padlock_sha_finup() 180 ctx->fallback = fallback_tfm; in padlock_init_tfm() 189 crypto_free_ahash(ctx->fallback); in padlock_exit_tfm()
|
| A D | sa2ul.c | 875 ctx->fallback.skcipher = child; in sa_cipher_cra_init() 1503 ctx->fallback.ahash = in sa_sha_cra_init_alg() 1505 if (IS_ERR(ctx->fallback.ahash)) { in sa_sha_cra_init_alg() 1508 return PTR_ERR(ctx->fallback.ahash); in sa_sha_cra_init_alg() 1657 crypto_free_ahash(ctx->fallback.ahash); in sa_sha_cra_exit() 1700 const char *fallback) in sa_cra_init_aead() argument 1715 ctx->fallback.aead = crypto_alloc_aead(fallback, 0, in sa_cra_init_aead() 1718 if (IS_ERR(ctx->fallback.aead)) { in sa_cra_init_aead() 1720 fallback); in sa_cra_init_aead() 1721 return PTR_ERR(ctx->fallback.aead); in sa_cra_init_aead() [all …]
|
| A D | img-hash.c | 114 struct crypto_ahash *fallback; member 493 ahash_request_set_tfm(&rctx->fallback_req, ctx->fallback); in img_hash_init() 558 ahash_request_set_tfm(&rctx->fallback_req, ctx->fallback); in img_hash_update() 573 ahash_request_set_tfm(&rctx->fallback_req, ctx->fallback); in img_hash_final() 588 ahash_request_set_tfm(&rctx->fallback_req, ctx->fallback); in img_hash_finup() 605 ahash_request_set_tfm(&rctx->fallback_req, ctx->fallback); in img_hash_import() 687 ctx->fallback = crypto_alloc_ahash(alg_name, 0, in img_hash_cra_init() 689 if (IS_ERR(ctx->fallback)) { in img_hash_cra_init() 691 return PTR_ERR(ctx->fallback); in img_hash_cra_init() 695 crypto_ahash_reqsize(ctx->fallback) + in img_hash_cra_init() [all …]
|
| A D | sahara.c | 143 struct crypto_skcipher *fallback; member 638 crypto_skcipher_clear_flags(ctx->fallback, CRYPTO_TFM_REQ_MASK); in sahara_aes_setkey() 639 crypto_skcipher_set_flags(ctx->fallback, tfm->base.crt_flags & in sahara_aes_setkey() 641 return crypto_skcipher_setkey(ctx->fallback, key, keylen); in sahara_aes_setkey() 650 skcipher_request_set_tfm(&rctx->fallback_req, ctx->fallback); in sahara_aes_fallback() 713 ctx->fallback = crypto_alloc_skcipher(name, 0, in sahara_aes_init_tfm() 715 if (IS_ERR(ctx->fallback)) { in sahara_aes_init_tfm() 717 return PTR_ERR(ctx->fallback); in sahara_aes_init_tfm() 721 crypto_skcipher_reqsize(ctx->fallback)); in sahara_aes_init_tfm() 730 crypto_free_skcipher(ctx->fallback); in sahara_aes_exit_tfm()
|
| A D | omap-aes.c | 543 skcipher_request_set_tfm(&rctx->fallback_req, ctx->fallback); in omap_aes_crypt() 583 crypto_skcipher_clear_flags(ctx->fallback, CRYPTO_TFM_REQ_MASK); in omap_aes_setkey() 584 crypto_skcipher_set_flags(ctx->fallback, tfm->base.crt_flags & in omap_aes_setkey() 587 ret = crypto_skcipher_setkey(ctx->fallback, key, keylen); in omap_aes_setkey() 634 ctx->fallback = blk; in omap_aes_init_tfm() 646 if (ctx->fallback) in omap_aes_exit_tfm() 647 crypto_free_skcipher(ctx->fallback); in omap_aes_exit_tfm() 649 ctx->fallback = NULL; in omap_aes_exit_tfm() 1081 static DEVICE_ATTR_RW(fallback);
|
| /drivers/tty/vt/ |
| A D | gen_ucs_fallback_table.py | 214 for code, fallback in fallback_map.items(): 216 if fallback == 0: 221 page_groups[page].append((offset, fallback)) 239 start_offset, fallback = entries[i] 245 entries[j][1] == fallback): # same fallback 255 compressed_entries.append((end_offset, fallback)) 334 for i, (offset, fallback) in enumerate(entries): 342 if fallback == RANGE_MARKER:
|
| A D | ucs.c | 173 u8 fallback; /* Fallback character or range start marker */ member 197 if (entry->fallback == UCS_PAGE_ENTRY_RANGE_MARKER) { in ucs_page_entry_cmp() 248 if (entry->fallback == UCS_PAGE_ENTRY_RANGE_MARKER) in ucs_get_fallback() 250 return entry->fallback; in ucs_get_fallback()
|
| /drivers/base/firmware_loader/ |
| A D | Kconfig | 95 bool "Enable the firmware sysfs fallback mechanism" 100 loading to the kernel through userspace as a fallback mechanism 106 refer to Documentation/driver-api/firmware/fallback-mechanisms.rst. 111 the requested firmware a sysfs fallback loading facility is made 114 this is known as the driver using the custom fallback mechanism. 115 If the custom fallback mechanism is used userspace must always 123 relies on or uses a fallback mechanism in userspace. If you need to 168 bool "Force the firmware sysfs fallback mechanism when possible" 171 Enabling this option forces a sysfs userspace fallback mechanism 173 a fallback mechanism. Firmware calls which do prohibit a fallback [all …]
|
| /drivers/gpu/drm/ |
| A D | drm_edid_load.c | 58 char *edidname, *last, *colon, *fwstr, *edidstr, *fallback = NULL; in drm_edid_load_firmware() local 87 fallback = edidname; in drm_edid_load_firmware() 91 if (!fallback) { in drm_edid_load_firmware() 95 edidname = fallback; in drm_edid_load_firmware()
|
| /drivers/gpu/drm/amd/amdgpu/ |
| A D | amdgpu_res_cursor.h | 63 goto fallback; in amdgpu_res_first() 77 goto fallback; in amdgpu_res_first() 104 goto fallback; in amdgpu_res_first() 109 fallback: in amdgpu_res_first()
|
| /drivers/crypto/ccp/ |
| A D | ccp-crypto-aes-xts.c | 111 unsigned int fallback = 0; in ccp_aes_xts_crypt() local 140 fallback = 1; in ccp_aes_xts_crypt() 143 fallback = 1; in ccp_aes_xts_crypt() 146 fallback = 1; in ccp_aes_xts_crypt() 147 if (fallback) { in ccp_aes_xts_crypt()
|
| /drivers/gpu/drm/vmwgfx/ |
| A D | vmwgfx_bo.c | 743 static u32 placement_flags(u32 domain, u32 desired, u32 fallback) in placement_flags() argument 745 if (desired & fallback & domain) in placement_flags() 755 set_placement_list(struct ttm_place *pl, u32 desired, u32 fallback) in set_placement_list() argument 757 u32 domain = desired | fallback; in set_placement_list() 766 fallback); in set_placement_list() 774 fallback); in set_placement_list() 782 fallback); in set_placement_list() 790 desired, fallback); in set_placement_list() 798 fallback); in set_placement_list()
|
| /drivers/crypto/qce/ |
| A D | skcipher.c | 205 ret = crypto_skcipher_setkey(ctx->fallback, key, keylen); in qce_skcipher_setkey() 295 skcipher_request_set_tfm(&rctx->fallback_req, ctx->fallback); in qce_skcipher_crypt() 332 ctx->fallback = crypto_alloc_skcipher(crypto_tfm_alg_name(&tfm->base), in qce_skcipher_init_fallback() 334 if (IS_ERR(ctx->fallback)) in qce_skcipher_init_fallback() 335 return PTR_ERR(ctx->fallback); in qce_skcipher_init_fallback() 338 crypto_skcipher_reqsize(ctx->fallback)); in qce_skcipher_init_fallback() 346 crypto_free_skcipher(ctx->fallback); in qce_skcipher_exit()
|
| A D | aead.c | 529 aead_request_set_tfm(&rctx->fallback_req, ctx->fallback); in qce_aead_crypt() 591 crypto_aead_setkey(ctx->fallback, key, keylen + QCE_CCM4309_SALT_SIZE) : in qce_aead_ccm_setkey() 592 crypto_aead_setkey(ctx->fallback, key, keylen); in qce_aead_ccm_setkey() 648 return crypto_aead_setkey(ctx->fallback, key, keylen); in qce_aead_setkey() 664 return crypto_aead_setauthsize(ctx->fallback, authsize); in qce_aead_setauthsize() 672 ctx->fallback = crypto_alloc_aead(crypto_tfm_alg_name(&tfm->base), in qce_aead_init() 675 if (IS_ERR(ctx->fallback)) in qce_aead_init() 676 return PTR_ERR(ctx->fallback); in qce_aead_init() 679 crypto_aead_reqsize(ctx->fallback)); in qce_aead_init() 687 crypto_free_aead(ctx->fallback); in qce_aead_exit()
|
| /drivers/gpu/drm/xe/ |
| A D | xe_res_cursor.h | 97 goto fallback; in xe_res_first() 117 goto fallback; in xe_res_first() 137 goto fallback; in xe_res_first() 142 fallback: in xe_res_first()
|
| /drivers/ufs/host/ |
| A D | ufs-renesas.c | 467 goto fallback; in ufs_renesas_init() 472 goto fallback; in ufs_renesas_init() 478 goto fallback; in ufs_renesas_init() 484 goto fallback; in ufs_renesas_init() 489 goto fallback; in ufs_renesas_init() 496 fallback: in ufs_renesas_init()
|
| /drivers/crypto/tegra/ |
| A D | tegra-se-hash.c | 27 bool fallback; member 358 if (ctx->fallback) in tegra_sha_do_init() 594 ctx->fallback = false; in tegra_sha_cra_init() 629 ctx->fallback = true; in tegra_hmac_fallback_setkey() 646 ctx->fallback = false; in tegra_hmac_setkey() 668 if (ctx->fallback) in tegra_sha_update() 682 if (ctx->fallback) in tegra_sha_final() 696 if (ctx->fallback) in tegra_sha_finup() 710 if (ctx->fallback) in tegra_sha_digest() 724 if (ctx->fallback) in tegra_sha_export() [all …]
|
| /drivers/cpufreq/ |
| A D | cpufreq-dt.c | 160 bool fallback = false; in dt_cpufreq_early_init() local 208 fallback = true; in dt_cpufreq_early_init() 240 if (fallback) { in dt_cpufreq_early_init()
|
| /drivers/gpu/drm/qxl/ |
| A D | qxl_object.c | 220 goto fallback; in qxl_bo_kmap_atomic_page() 224 fallback: in qxl_bo_kmap_atomic_page() 271 goto fallback; in qxl_bo_kunmap_atomic_page() 275 fallback: in qxl_bo_kunmap_atomic_page()
|
| /drivers/pci/hotplug/ |
| A D | rpaphp_pci.c | 90 goto fallback; in __rpaphp_get_sensor_state() 98 fallback: in __rpaphp_get_sensor_state()
|
| /drivers/crypto/caam/ |
| A D | caamalg_qi.c | 78 struct crypto_skcipher *fallback; member 1433 if (!req->cryptlen && !ctx->fallback) in skcipher_crypt() 2516 struct crypto_skcipher *fallback; in caam_cra_init() local 2520 if (IS_ERR(fallback)) { in caam_cra_init() 2522 tfm_name, PTR_ERR(fallback)); in caam_cra_init() 2523 return PTR_ERR(fallback); in caam_cra_init() 2526 ctx->fallback = fallback; in caam_cra_init() 2532 if (ret && ctx->fallback) in caam_cra_init() 2533 crypto_free_skcipher(ctx->fallback); in caam_cra_init() 2563 if (ctx->fallback) in caam_cra_exit() [all …]
|
| /drivers/gpu/drm/omapdrm/ |
| A D | omap_fbdev.c | 65 goto fallback; in FB_GEN_DEFAULT_DEFERRED_DMAMEM_OPS() 71 goto fallback; in FB_GEN_DEFAULT_DEFERRED_DMAMEM_OPS() 80 fallback: in FB_GEN_DEFAULT_DEFERRED_DMAMEM_OPS()
|