| /kernel/linux/linux-6.6/security/landlock/ |
| D | ruleset.c | 73 const struct landlock_layer (*const layers)[], const u32 num_layers, in create_rule() argument 88 new_rule = kzalloc(struct_size(new_rule, layers, new_num_layers), in create_rule() 97 memcpy(new_rule->layers, layers, in create_rule() 98 flex_array_size(new_rule, layers, num_layers)); in create_rule() 101 new_rule->layers[new_rule->num_layers - 1] = *new_layer; in create_rule() 133 * @layers: One or multiple layers to be copied into the new rule. 134 * @num_layers: The number of @layers entries. 136 * When user space requests to add a new rule to a ruleset, @layers only 141 * When merging a ruleset in a domain, or copying a domain, @layers will be 147 const struct landlock_layer (*const layers)[], in insert_rule() argument [all …]
|
| D | ruleset.h | 29 /* Makes sure all layers can be checked. */ 63 * @num_layers: Number of entries in @layers. 67 * @layers: Stack of layers, from the latest to the newest, implemented 70 struct landlock_layer layers[] __counted_by(num_layers); 133 * @num_layers: Number of layers that are used in this 134 * ruleset. This enables to check that all the layers 142 * saves all layers of merged rulesets in a stack 144 * one. These layers are used when merging rulesets, 148 * layers are set once and never changed for the
|
| /kernel/linux/linux-6.6/Documentation/devicetree/bindings/display/ |
| D | xylon,logicvc-display.yaml | 14 The Xylon LogiCVC is a display controller that supports multiple layers. 24 Layers are declared in the "layers" sub-node and have dedicated configuration. 109 xylon,layers-configurable: 112 Configuration of layers' size, position and offset is enabled 115 layers: 187 The description of the display controller layers, containing layer 207 - layers 238 xylon,layers-configurable; 240 layers {
|
| /kernel/linux/linux-5.10/drivers/edac/ |
| D | pasemi_edac.c | 183 struct edac_mc_layer layers[2]; in pasemi_edac_probe() local 200 layers[0].type = EDAC_MC_LAYER_CHIP_SELECT; in pasemi_edac_probe() 201 layers[0].size = PASEMI_EDAC_NR_CSROWS; in pasemi_edac_probe() 202 layers[0].is_virt_csrow = true; in pasemi_edac_probe() 203 layers[1].type = EDAC_MC_LAYER_CHANNEL; in pasemi_edac_probe() 204 layers[1].size = PASEMI_EDAC_NR_CHANS; in pasemi_edac_probe() 205 layers[1].is_virt_csrow = false; in pasemi_edac_probe() 206 mci = edac_mc_alloc(system_mmc_id++, ARRAY_SIZE(layers), layers, in pasemi_edac_probe()
|
| D | highbank_mc_edac.c | 148 struct edac_mc_layer layers[2]; in highbank_mc_probe() local 162 layers[0].type = EDAC_MC_LAYER_CHIP_SELECT; in highbank_mc_probe() 163 layers[0].size = 1; in highbank_mc_probe() 164 layers[0].is_virt_csrow = true; in highbank_mc_probe() 165 layers[1].type = EDAC_MC_LAYER_CHANNEL; in highbank_mc_probe() 166 layers[1].size = 1; in highbank_mc_probe() 167 layers[1].is_virt_csrow = false; in highbank_mc_probe() 168 mci = edac_mc_alloc(0, ARRAY_SIZE(layers), layers, in highbank_mc_probe()
|
| D | cell_edac.c | 172 struct edac_mc_layer layers[2]; in cell_edac_probe() local 202 layers[0].type = EDAC_MC_LAYER_CHIP_SELECT; in cell_edac_probe() 203 layers[0].size = 1; in cell_edac_probe() 204 layers[0].is_virt_csrow = true; in cell_edac_probe() 205 layers[1].type = EDAC_MC_LAYER_CHANNEL; in cell_edac_probe() 206 layers[1].size = num_chans; in cell_edac_probe() 207 layers[1].is_virt_csrow = false; in cell_edac_probe() 208 mci = edac_mc_alloc(pdev->id, ARRAY_SIZE(layers), layers, in cell_edac_probe()
|
| D | amd76x_edac.c | 238 struct edac_mc_layer layers[2]; in amd76x_probe1() local 247 layers[0].type = EDAC_MC_LAYER_CHIP_SELECT; in amd76x_probe1() 248 layers[0].size = AMD76X_NR_CSROWS; in amd76x_probe1() 249 layers[0].is_virt_csrow = true; in amd76x_probe1() 250 layers[1].type = EDAC_MC_LAYER_CHANNEL; in amd76x_probe1() 251 layers[1].size = 1; in amd76x_probe1() 252 layers[1].is_virt_csrow = false; in amd76x_probe1() 253 mci = edac_mc_alloc(0, ARRAY_SIZE(layers), layers, 0); in amd76x_probe1()
|
| D | i82860_edac.c | 188 struct edac_mc_layer layers[2]; in i82860_probe1() local 201 layers[0].type = EDAC_MC_LAYER_CHANNEL; in i82860_probe1() 202 layers[0].size = 2; in i82860_probe1() 203 layers[0].is_virt_csrow = true; in i82860_probe1() 204 layers[1].type = EDAC_MC_LAYER_SLOT; in i82860_probe1() 205 layers[1].size = 8; in i82860_probe1() 206 layers[1].is_virt_csrow = true; in i82860_probe1() 207 mci = edac_mc_alloc(0, ARRAY_SIZE(layers), layers, 0); in i82860_probe1()
|
| D | aspeed_edac.c | 282 struct edac_mc_layer layers[2]; in aspeed_probe() local 307 layers[0].type = EDAC_MC_LAYER_CHIP_SELECT; in aspeed_probe() 308 layers[0].size = 1; in aspeed_probe() 309 layers[0].is_virt_csrow = true; in aspeed_probe() 310 layers[1].type = EDAC_MC_LAYER_CHANNEL; in aspeed_probe() 311 layers[1].size = 1; in aspeed_probe() 312 layers[1].is_virt_csrow = false; in aspeed_probe() 314 mci = edac_mc_alloc(0, ARRAY_SIZE(layers), layers, 0); in aspeed_probe()
|
| D | octeon_edac-lmc.c | 228 struct edac_mc_layer layers[1]; in octeon_lmc_edac_probe() local 233 layers[0].type = EDAC_MC_LAYER_CHANNEL; in octeon_lmc_edac_probe() 234 layers[0].size = 1; in octeon_lmc_edac_probe() 235 layers[0].is_virt_csrow = false; in octeon_lmc_edac_probe() 246 mci = edac_mc_alloc(mc, ARRAY_SIZE(layers), layers, sizeof(struct octeon_lmc_pvt)); in octeon_lmc_edac_probe() 278 mci = edac_mc_alloc(mc, ARRAY_SIZE(layers), layers, sizeof(struct octeon_lmc_pvt)); in octeon_lmc_edac_probe()
|
| /kernel/linux/linux-6.6/drivers/edac/ |
| D | pasemi_edac.c | 183 struct edac_mc_layer layers[2]; in pasemi_edac_probe() local 200 layers[0].type = EDAC_MC_LAYER_CHIP_SELECT; in pasemi_edac_probe() 201 layers[0].size = PASEMI_EDAC_NR_CSROWS; in pasemi_edac_probe() 202 layers[0].is_virt_csrow = true; in pasemi_edac_probe() 203 layers[1].type = EDAC_MC_LAYER_CHANNEL; in pasemi_edac_probe() 204 layers[1].size = PASEMI_EDAC_NR_CHANS; in pasemi_edac_probe() 205 layers[1].is_virt_csrow = false; in pasemi_edac_probe() 206 mci = edac_mc_alloc(system_mmc_id++, ARRAY_SIZE(layers), layers, in pasemi_edac_probe()
|
| D | highbank_mc_edac.c | 149 struct edac_mc_layer layers[2]; in highbank_mc_probe() local 163 layers[0].type = EDAC_MC_LAYER_CHIP_SELECT; in highbank_mc_probe() 164 layers[0].size = 1; in highbank_mc_probe() 165 layers[0].is_virt_csrow = true; in highbank_mc_probe() 166 layers[1].type = EDAC_MC_LAYER_CHANNEL; in highbank_mc_probe() 167 layers[1].size = 1; in highbank_mc_probe() 168 layers[1].is_virt_csrow = false; in highbank_mc_probe() 169 mci = edac_mc_alloc(0, ARRAY_SIZE(layers), layers, in highbank_mc_probe()
|
| D | cell_edac.c | 172 struct edac_mc_layer layers[2]; in cell_edac_probe() local 202 layers[0].type = EDAC_MC_LAYER_CHIP_SELECT; in cell_edac_probe() 203 layers[0].size = 1; in cell_edac_probe() 204 layers[0].is_virt_csrow = true; in cell_edac_probe() 205 layers[1].type = EDAC_MC_LAYER_CHANNEL; in cell_edac_probe() 206 layers[1].size = num_chans; in cell_edac_probe() 207 layers[1].is_virt_csrow = false; in cell_edac_probe() 208 mci = edac_mc_alloc(pdev->id, ARRAY_SIZE(layers), layers, in cell_edac_probe()
|
| D | amd76x_edac.c | 237 struct edac_mc_layer layers[2]; in amd76x_probe1() local 246 layers[0].type = EDAC_MC_LAYER_CHIP_SELECT; in amd76x_probe1() 247 layers[0].size = AMD76X_NR_CSROWS; in amd76x_probe1() 248 layers[0].is_virt_csrow = true; in amd76x_probe1() 249 layers[1].type = EDAC_MC_LAYER_CHANNEL; in amd76x_probe1() 250 layers[1].size = 1; in amd76x_probe1() 251 layers[1].is_virt_csrow = false; in amd76x_probe1() 252 mci = edac_mc_alloc(0, ARRAY_SIZE(layers), layers, 0); in amd76x_probe1()
|
| D | aspeed_edac.c | 282 struct edac_mc_layer layers[2]; in aspeed_probe() local 307 layers[0].type = EDAC_MC_LAYER_CHIP_SELECT; in aspeed_probe() 308 layers[0].size = 1; in aspeed_probe() 309 layers[0].is_virt_csrow = true; in aspeed_probe() 310 layers[1].type = EDAC_MC_LAYER_CHANNEL; in aspeed_probe() 311 layers[1].size = 1; in aspeed_probe() 312 layers[1].is_virt_csrow = false; in aspeed_probe() 314 mci = edac_mc_alloc(0, ARRAY_SIZE(layers), layers, 0); in aspeed_probe()
|
| D | i82860_edac.c | 187 struct edac_mc_layer layers[2]; in i82860_probe1() local 200 layers[0].type = EDAC_MC_LAYER_CHANNEL; in i82860_probe1() 201 layers[0].size = 2; in i82860_probe1() 202 layers[0].is_virt_csrow = true; in i82860_probe1() 203 layers[1].type = EDAC_MC_LAYER_SLOT; in i82860_probe1() 204 layers[1].size = 8; in i82860_probe1() 205 layers[1].is_virt_csrow = true; in i82860_probe1() 206 mci = edac_mc_alloc(0, ARRAY_SIZE(layers), layers, 0); in i82860_probe1()
|
| D | octeon_edac-lmc.c | 228 struct edac_mc_layer layers[1]; in octeon_lmc_edac_probe() local 233 layers[0].type = EDAC_MC_LAYER_CHANNEL; in octeon_lmc_edac_probe() 234 layers[0].size = 1; in octeon_lmc_edac_probe() 235 layers[0].is_virt_csrow = false; in octeon_lmc_edac_probe() 246 mci = edac_mc_alloc(mc, ARRAY_SIZE(layers), layers, sizeof(struct octeon_lmc_pvt)); in octeon_lmc_edac_probe() 278 mci = edac_mc_alloc(mc, ARRAY_SIZE(layers), layers, sizeof(struct octeon_lmc_pvt)); in octeon_lmc_edac_probe()
|
| /kernel/linux/linux-5.10/drivers/media/dvb-frontends/ |
| D | tc90522.c | 201 int layers; in tc90522s_get_frontend() local 209 layers = 0; in tc90522s_get_frontend() 236 layers = (v > 0) ? 2 : 1; in tc90522s_get_frontend() 284 stats->len = layers; in tc90522s_get_frontend() 287 for (i = 0; i < layers; i++) in tc90522s_get_frontend() 290 for (i = 0; i < layers; i++) { in tc90522s_get_frontend() 298 stats->len = layers; in tc90522s_get_frontend() 300 for (i = 0; i < layers; i++) in tc90522s_get_frontend() 303 for (i = 0; i < layers; i++) { in tc90522s_get_frontend() 336 int layers; in tc90522t_get_frontend() local [all …]
|
| /kernel/linux/linux-6.6/drivers/media/dvb-frontends/ |
| D | tc90522.c | 201 int layers; in tc90522s_get_frontend() local 209 layers = 0; in tc90522s_get_frontend() 236 layers = (v > 0) ? 2 : 1; in tc90522s_get_frontend() 284 stats->len = layers; in tc90522s_get_frontend() 287 for (i = 0; i < layers; i++) in tc90522s_get_frontend() 290 for (i = 0; i < layers; i++) { in tc90522s_get_frontend() 298 stats->len = layers; in tc90522s_get_frontend() 300 for (i = 0; i < layers; i++) in tc90522s_get_frontend() 303 for (i = 0; i < layers; i++) { in tc90522s_get_frontend() 336 int layers; in tc90522t_get_frontend() local [all …]
|
| /kernel/linux/linux-6.6/Documentation/filesystems/ |
| D | overlayfs.rst | 32 In the special case of all overlay layers on the same underlying 39 On 64bit systems, even if all overlay layers are not on the same 66 | All layers | Y | Y | Y | Y | Y | Y | Y | Y | 69 | Layers not | N | N | Y | N | N | Y | N | Y | 308 Check (b) ensures that no task gains permissions to underlying layers that 327 Multiple lower layers 330 Multiple lower layers can now be given using the colon (":") as a 378 for untrusted layers like from a pen drive. 387 Data-only lower layers 391 of information from up to three different layers: [all …]
|
| /kernel/linux/linux-6.6/fs/overlayfs/ |
| D | ovl_entry.h | 58 /* Number of unique fs among layers including upper fs */ 60 /* Number of data-only lower layers */ 62 const struct ovl_layer *layers; member 96 /* Number of lower layers, not including data-only layers */ 104 return ofs->layers[0].mnt; in ovl_upper_mnt()
|
| D | super.c | 399 * file handles, so they require that all layers support them. in ovl_lower_dir() 548 pr_err("upper fs is r/o, try multi-lower layers mount\n"); in ovl_get_upper() 967 * as all lower layers with null uuid are on the same fs. in ovl_lower_uuid_ok() 1028 * The fsid after the last lower fsid is used for the data layers. 1038 struct ovl_fs_context *ctx, struct ovl_layer *layers) in ovl_get_layers() argument 1050 * and the last fsid is reserved for "null fs" of the data layers. in ovl_get_layers() 1055 * All lower layers that share the same fs as upper layer, use the same in ovl_get_layers() 1086 * Check if lower root conflicts with this overlay layers before in ovl_get_layers() 1113 * Make lower layers R/O. That way fchmod/fchown on lower file in ovl_get_layers() 1118 layers[ofs->numlayer].trap = trap; in ovl_get_layers() [all …]
|
| /kernel/linux/linux-5.10/Documentation/block/ |
| D | inline-encryption.rst | 44 - We need a way for upper layers like filesystems to specify an encryption 49 capabilities in a unified way to the upper layers. 57 encryption context from the upper layers (like the fs layer) to the 67 upper layers. The generic mode of operation is: each device driver that wants 69 Upper layers that want to use IE on this device can then use this KSM in 93 We introduce ``block/blk-crypto-fallback.c``, which allows upper layers to remain 149 ``blk_crypto_init_key`` allows upper layers to initialize such a 158 ``blk_crypto_config_supported`` allows upper layers to query whether or not the 166 ``blk_crypto_start_using_key`` - Upper layers must call this function on 174 ``blk_crypto_evict_key`` *must* be called by upper layers before a
|
| /kernel/linux/linux-6.6/Documentation/scsi/ |
| D | scsi_eh.rst | 152 Note that this does not mean lower layers are quiescent. If a LLDD 153 completed a scmd with error status, the LLDD and lower layers are 155 has timed out, unless hostt->eh_timed_out() made lower layers forget 157 active as long as lower layers are concerned and completion could 206 lower layers and lower layers are ready to process or fail the scmd 389 that lower layers have forgotten about the scmd and we can 398 and STU doesn't make lower layers forget about those 400 if STU succeeds leaving lower layers in an inconsistent 453 On completion, the handler should have made lower layers forget about 496 - Know that timed out scmds are still active on lower layers. Make [all …]
|
| /kernel/linux/linux-5.10/Documentation/scsi/ |
| D | scsi_eh.rst | 151 Note that this does not mean lower layers are quiescent. If a LLDD 152 completed a scmd with error status, the LLDD and lower layers are 154 has timed out, unless hostt->eh_timed_out() made lower layers forget 156 active as long as lower layers are concerned and completion could 205 lower layers and lower layers are ready to process or fail the scmd 388 that lower layers have forgotten about the scmd and we can 397 and STU doesn't make lower layers forget about those 399 if STU succeeds leaving lower layers in an inconsistent 452 On completion, the handler should have made lower layers forget about 495 - Know that timed out scmds are still active on lower layers. Make [all …]
|