Home
last modified time | relevance | path

Searched full:layers (Results 1 – 25 of 491) sorted by relevance

12345678910>>...20

/linux/security/landlock/
H A Druleset.c103 const struct landlock_layer (*const layers)[], const u32 num_layers, in create_rule() argument
118 new_rule = kzalloc(struct_size(new_rule, layers, new_num_layers), in create_rule()
132 memcpy(new_rule->layers, layers, in create_rule()
133 flex_array_size(new_rule, layers, num_layers)); in create_rule()
136 new_rule->layers[new_rule->num_layers - 1] = *new_layer; in create_rule()
186 * @layers: One or multiple layers to be copied into the new rule.
187 * @num_layers: The number of @layers entries.
189 * When user space requests to add a new rule to a ruleset, @layers only
194 * When merging a ruleset in a domain, or copying a domain, @layers will be
200 const struct landlock_layer (*const layers)[], in insert_rule() argument
[all …]
H A Druleset.h61 /* Makes sure all layers can be checked. */
142 * @num_layers: Number of entries in @layers.
146 * @layers: Stack of layers, from the latest to the newest, implemented
149 struct landlock_layer layers[] __counted_by(num_layers);
224 * @num_layers: Number of layers that are used in this
225 * ruleset. This enables to check that all the layers
233 * A domain saves all layers of merged rulesets in a
235 * last one. These layers are used when merging
239 * layers are set once and never changed for the
/linux/Documentation/devicetree/bindings/display/
H A Dxylon,logicvc-display.yaml14 The Xylon LogiCVC is a display controller that supports multiple layers.
24 Layers are declared in the "layers" sub-node and have dedicated configuration.
109 xylon,layers-configurable:
112 Configuration of layers' size, position and offset is enabled
115 layers:
187 The description of the display controller layers, containing layer
207 - layers
238 xylon,layers-configurable;
240 layers {
/linux/drivers/edac/
H A Dpasemi_edac.c183 struct edac_mc_layer layers[2]; in pasemi_edac_probe() local
200 layers[0].type = EDAC_MC_LAYER_CHIP_SELECT; in pasemi_edac_probe()
201 layers[0].size = PASEMI_EDAC_NR_CSROWS; in pasemi_edac_probe()
202 layers[0].is_virt_csrow = true; in pasemi_edac_probe()
203 layers[1].type = EDAC_MC_LAYER_CHANNEL; in pasemi_edac_probe()
204 layers[1].size = PASEMI_EDAC_NR_CHANS; in pasemi_edac_probe()
205 layers[1].is_virt_csrow = false; in pasemi_edac_probe()
206 mci = edac_mc_alloc(system_mmc_id++, ARRAY_SIZE(layers), layers, in pasemi_edac_probe()
H A Dhighbank_mc_edac.c149 struct edac_mc_layer layers[2]; in highbank_mc_probe() local
163 layers[0].type = EDAC_MC_LAYER_CHIP_SELECT; in highbank_mc_probe()
164 layers[0].size = 1; in highbank_mc_probe()
165 layers[0].is_virt_csrow = true; in highbank_mc_probe()
166 layers[1].type = EDAC_MC_LAYER_CHANNEL; in highbank_mc_probe()
167 layers[1].size = 1; in highbank_mc_probe()
168 layers[1].is_virt_csrow = false; in highbank_mc_probe()
169 mci = edac_mc_alloc(0, ARRAY_SIZE(layers), layers, in highbank_mc_probe()
H A Dcell_edac.c172 struct edac_mc_layer layers[2]; in cell_edac_probe() local
202 layers[0].type = EDAC_MC_LAYER_CHIP_SELECT; in cell_edac_probe()
203 layers[0].size = 1; in cell_edac_probe()
204 layers[0].is_virt_csrow = true; in cell_edac_probe()
205 layers[1].type = EDAC_MC_LAYER_CHANNEL; in cell_edac_probe()
206 layers[1].size = num_chans; in cell_edac_probe()
207 layers[1].is_virt_csrow = false; in cell_edac_probe()
208 mci = edac_mc_alloc(pdev->id, ARRAY_SIZE(layers), layers, in cell_edac_probe()
H A Di82860_edac.c187 struct edac_mc_layer layers[2]; in i82860_probe1() local
200 layers[0].type = EDAC_MC_LAYER_CHANNEL; in i82860_probe1()
201 layers[0].size = 2; in i82860_probe1()
202 layers[0].is_virt_csrow = true; in i82860_probe1()
203 layers[1].type = EDAC_MC_LAYER_SLOT; in i82860_probe1()
204 layers[1].size = 8; in i82860_probe1()
205 layers[1].is_virt_csrow = true; in i82860_probe1()
206 mci = edac_mc_alloc(0, ARRAY_SIZE(layers), layers, 0); in i82860_probe1()
H A Damd76x_edac.c237 struct edac_mc_layer layers[2]; in amd76x_probe1() local
246 layers[0].type = EDAC_MC_LAYER_CHIP_SELECT; in amd76x_probe1()
247 layers[0].size = AMD76X_NR_CSROWS; in amd76x_probe1()
248 layers[0].is_virt_csrow = true; in amd76x_probe1()
249 layers[1].type = EDAC_MC_LAYER_CHANNEL; in amd76x_probe1()
250 layers[1].size = 1; in amd76x_probe1()
251 layers[1].is_virt_csrow = false; in amd76x_probe1()
252 mci = edac_mc_alloc(0, ARRAY_SIZE(layers), layers, 0); in amd76x_probe1()
H A Daspeed_edac.c282 struct edac_mc_layer layers[2]; in aspeed_probe() local
307 layers[0].type = EDAC_MC_LAYER_CHIP_SELECT; in aspeed_probe()
308 layers[0].size = 1; in aspeed_probe()
309 layers[0].is_virt_csrow = true; in aspeed_probe()
310 layers[1].type = EDAC_MC_LAYER_CHANNEL; in aspeed_probe()
311 layers[1].size = 1; in aspeed_probe()
312 layers[1].is_virt_csrow = false; in aspeed_probe()
314 mci = edac_mc_alloc(0, ARRAY_SIZE(layers), layers, 0); in aspeed_probe()
H A Docteon_edac-lmc.c228 struct edac_mc_layer layers[1]; in octeon_lmc_edac_probe() local
233 layers[0].type = EDAC_MC_LAYER_CHANNEL; in octeon_lmc_edac_probe()
234 layers[0].size = 1; in octeon_lmc_edac_probe()
235 layers[0].is_virt_csrow = false; in octeon_lmc_edac_probe()
246 mci = edac_mc_alloc(mc, ARRAY_SIZE(layers), layers, sizeof(struct octeon_lmc_pvt)); in octeon_lmc_edac_probe()
278 mci = edac_mc_alloc(mc, ARRAY_SIZE(layers), layers, sizeof(struct octeon_lmc_pvt)); in octeon_lmc_edac_probe()
H A Dx38_edac.c322 struct edac_mc_layer layers[2]; in x38_probe1() local
338 layers[0].type = EDAC_MC_LAYER_CHIP_SELECT; in x38_probe1()
339 layers[0].size = X38_RANKS; in x38_probe1()
340 layers[0].is_virt_csrow = true; in x38_probe1()
341 layers[1].type = EDAC_MC_LAYER_CHANNEL; in x38_probe1()
342 layers[1].size = x38_channel_num; in x38_probe1()
343 layers[1].is_virt_csrow = false; in x38_probe1()
344 mci = edac_mc_alloc(0, ARRAY_SIZE(layers), layers, 0); in x38_probe1()
H A Dr82600_edac.c271 struct edac_mc_layer layers[2]; in r82600_probe1() local
285 layers[0].type = EDAC_MC_LAYER_CHIP_SELECT; in r82600_probe1()
286 layers[0].size = R82600_NR_CSROWS; in r82600_probe1()
287 layers[0].is_virt_csrow = true; in r82600_probe1()
288 layers[1].type = EDAC_MC_LAYER_CHANNEL; in r82600_probe1()
289 layers[1].size = R82600_NR_CHANS; in r82600_probe1()
290 layers[1].is_virt_csrow = false; in r82600_probe1()
291 mci = edac_mc_alloc(0, ARRAY_SIZE(layers), layers, 0); in r82600_probe1()
H A Dedac_mc.c70 edac_layer_name[mci->layers[i].type], in edac_dimm_info_location()
206 kfree(mci->layers); in mci_release()
296 edac_layer_name[mci->layers[layer].type], in edac_mc_alloc_dimms()
309 if (mci->layers[0].is_virt_csrow) { in edac_mc_alloc_dimms()
326 if (pos[layer] < mci->layers[layer].size) in edac_mc_alloc_dimms()
337 struct edac_mc_layer *layers, in edac_mc_alloc() argument
354 tot_dimms *= layers[idx].size; in edac_mc_alloc()
356 if (layers[idx].is_virt_csrow) in edac_mc_alloc()
357 tot_csrows *= layers[idx].size; in edac_mc_alloc()
359 tot_channels *= layers[idx].size; in edac_mc_alloc()
[all …]
H A Di82443bxgx_edac.c234 struct edac_mc_layer layers[2]; in i82443bxgx_edacmc_probe1() local
248 layers[0].type = EDAC_MC_LAYER_CHIP_SELECT; in i82443bxgx_edacmc_probe1()
249 layers[0].size = I82443BXGX_NR_CSROWS; in i82443bxgx_edacmc_probe1()
250 layers[0].is_virt_csrow = true; in i82443bxgx_edacmc_probe1()
251 layers[1].type = EDAC_MC_LAYER_CHANNEL; in i82443bxgx_edacmc_probe1()
252 layers[1].size = I82443BXGX_NR_CHANS; in i82443bxgx_edacmc_probe1()
253 layers[1].is_virt_csrow = false; in i82443bxgx_edacmc_probe1()
254 mci = edac_mc_alloc(0, ARRAY_SIZE(layers), layers, 0); in i82443bxgx_edacmc_probe1()
H A Di3200_edac.c340 struct edac_mc_layer layers[2]; in i3200_probe1() local
355 layers[0].type = EDAC_MC_LAYER_CHIP_SELECT; in i3200_probe1()
356 layers[0].size = I3200_DIMMS; in i3200_probe1()
357 layers[0].is_virt_csrow = true; in i3200_probe1()
358 layers[1].type = EDAC_MC_LAYER_CHANNEL; in i3200_probe1()
359 layers[1].size = nr_channels; in i3200_probe1()
360 layers[1].is_virt_csrow = false; in i3200_probe1()
361 mci = edac_mc_alloc(0, ARRAY_SIZE(layers), layers, in i3200_probe1()
H A Di3000_edac.c313 struct edac_mc_layer layers[2]; in i3000_probe1() local
356 layers[0].type = EDAC_MC_LAYER_CHIP_SELECT; in i3000_probe1()
357 layers[0].size = I3000_RANKS / nr_channels; in i3000_probe1()
358 layers[0].is_virt_csrow = true; in i3000_probe1()
359 layers[1].type = EDAC_MC_LAYER_CHANNEL; in i3000_probe1()
360 layers[1].size = nr_channels; in i3000_probe1()
361 layers[1].is_virt_csrow = false; in i3000_probe1()
362 mci = edac_mc_alloc(0, ARRAY_SIZE(layers), layers, 0); in i3000_probe1()
/linux/drivers/media/dvb-frontends/
H A Dtc90522.c201 int layers; in tc90522s_get_frontend() local
209 layers = 0; in tc90522s_get_frontend()
236 layers = (v > 0) ? 2 : 1; in tc90522s_get_frontend()
284 stats->len = layers; in tc90522s_get_frontend()
287 for (i = 0; i < layers; i++) in tc90522s_get_frontend()
290 for (i = 0; i < layers; i++) { in tc90522s_get_frontend()
298 stats->len = layers; in tc90522s_get_frontend()
300 for (i = 0; i < layers; i++) in tc90522s_get_frontend()
303 for (i = 0; i < layers; i++) { in tc90522s_get_frontend()
336 int layers; in tc90522t_get_frontend() local
[all …]
/linux/Documentation/filesystems/
H A Doverlayfs.rst32 In the special case of all overlay layers on the same underlying
39 On 64bit systems, even if all overlay layers are not on the same
66 | All layers | Y | Y | Y | Y | Y | Y | Y | Y |
69 | Layers not | N | N | Y | N | N | Y | N | Y |
317 Check (b) ensures that no task gains permissions to underlying layers that
336 Multiple lower layers
339 Multiple lower layers can now be given using the colon (":") as a
390 for untrusted layers like from a pen drive.
399 Data-only lower layers
403 of information from up to three different layers:
[all …]
/linux/fs/overlayfs/
H A Dsuper.c393 * file handles, so they require that all layers support them. in ovl_lower_dir()
482 pr_err("upper fs is r/o, try multi-lower layers mount\n"); in ovl_get_upper()
906 * as all lower layers with null uuid are on the same fs. in ovl_lower_uuid_ok()
967 * The fsid after the last lower fsid is used for the data layers.
977 struct ovl_fs_context *ctx, struct ovl_layer *layers) in ovl_get_layers() argument
989 * and the last fsid is reserved for "null fs" of the data layers. in ovl_get_layers()
994 * All lower layers that share the same fs as upper layer, use the same in ovl_get_layers()
1025 * Check if lower root conflicts with this overlay layers before in ovl_get_layers()
1052 * Make lower layers R/O. That way fchmod/fchown on lower file in ovl_get_layers()
1057 layers[ofs->numlayer].trap = trap; in ovl_get_layers()
[all …]
H A Dovl_entry.h60 /* Number of unique fs among layers including upper fs */
62 /* Number of data-only lower layers */
64 struct ovl_layer *layers; member
95 /* Number of lower layers, not including data-only layers */
103 return ofs->layers[0].mnt; in ovl_upper_mnt()
/linux/drivers/gpu/drm/atmel-hlcdc/
H A Datmel_hlcdc_dc.h177 * can be placed differently on 2 different layers depending on its
343 * @layers: active HLCDC layers
353 struct atmel_hlcdc_layer *layers[ATMEL_HLCDC_MAX_LAYERS]; member
366 * @update_lcdc_buffers: update the each LCDC layers DMA registers
367 * @lcdc_atomic_disable: disable LCDC interrupts and layers
368 * @lcdc_update_general_settings: update each LCDC layers general
370 * @lcdc_atomic_update: enable the LCDC layers and interrupts
412 * @layers: a layer description table describing available layers
427 const struct atmel_hlcdc_layer_desc *layers; member
/linux/Documentation/scsi/
H A Dscsi_eh.rst152 Note that this does not mean lower layers are quiescent. If a LLDD
153 completed a scmd with error status, the LLDD and lower layers are
155 has timed out, unless hostt->eh_timed_out() made lower layers forget
157 active as long as lower layers are concerned and completion could
206 lower layers and lower layers are ready to process or fail the scmd
389 that lower layers have forgotten about the scmd and we can
398 and STU doesn't make lower layers forget about those
400 if STU succeeds leaving lower layers in an inconsistent
453 On completion, the handler should have made lower layers forget about
496 - Know that timed out scmds are still active on lower layers. Make
[all …]
/linux/Documentation/block/
H A Dinline-encryption.rst50 - We need a way for upper layers (e.g. filesystems) to specify an encryption
59 advertise crypto capabilities to upper layers in a generic way.
67 - Upper layers typically define a specific end-of-life for crypto keys, e.g.
70 layers to also evict keys from any keyslots they are present in.
98 functions to program and evict keys) to upper layers. Each device driver that
135 It is desirable for the inline encryption support of upper layers (e.g.
138 to allow upper layers to just always use inline encryption rather than have to
164 encryption context. Therefore, lower layers only see standard unencrypted I/O.
240 blk_crypto_profile to tell upper layers how to control the inline encryption
/linux/drivers/gpu/drm/xlnx/
H A Dzynqmp_dpsub.h56 * @layers: Video and graphics layers
76 struct zynqmp_disp_layer *layers[ZYNQMP_DPSUB_NUM_LAYERS]; member
/linux/drivers/staging/most/Documentation/
H A Ddriver_usage.txt8 MOST defines the protocol, hardware and software layers necessary to allow
19 consumer devices via optical or electrical physical layers directly to one
27 three layers. From bottom up these layers are: the adapter layer, the core
31 routing through all three layers, the configuration of the driver, the
35 For each of the other two layers a set of modules is provided. Those can be

12345678910>>...20