/linux/fs/minix/ |
H A D | itree_common.c | 158 Indirect *partial; in get_block() local 166 partial = get_branch(inode, depth, offsets, chain, &err); in get_block() 169 if (!partial) { in get_block() 173 partial = chain+depth-1; /* the whole chain */ in get_block() 180 while (partial > chain) { in get_block() 181 brelse(partial->bh); in get_block() 182 partial--; in get_block() 196 left = (chain + depth) - partial; in get_block() 197 err = alloc_branch(inode, left, offsets+(partial-chain), partial); in get_block() 201 if (splice_branch(inode, chain, partial, left) < 0) in get_block() [all …]
|
/linux/fs/ext4/ |
H A D | indirect.c | 244 * @partial: pointer to the last triple within a chain 252 Indirect *partial) in ext4_find_goal() argument 260 goal = ext4_find_near(inode, partial); in ext4_find_goal() 316 * we had read the existing part of chain and partial points to the last 538 Indirect *partial; in ext4_ind_map_blocks() local 554 partial = ext4_get_branch(inode, depth, offsets, chain, &err); in ext4_ind_map_blocks() 557 if (!partial) { in ext4_ind_map_blocks() 580 * Count number blocks in a subtree under 'partial'. At each in ext4_ind_map_blocks() 586 for (i = partial - chain + 1; i < depth; i++) in ext4_ind_map_blocks() 620 ar.goal = ext4_find_goal(inode, map->m_lblk, partial); in ext4_ind_map_blocks() [all …]
|
/linux/lib/crypto/ |
H A D | sha1.c | 171 size_t partial = ctx->bytecount % SHA1_BLOCK_SIZE; in sha1_update() local 175 if (partial + len >= SHA1_BLOCK_SIZE) { in sha1_update() 178 if (partial) { in sha1_update() 179 size_t l = SHA1_BLOCK_SIZE - partial; in sha1_update() 181 memcpy(&ctx->buf[partial], data, l); in sha1_update() 195 partial = 0; in sha1_update() 198 memcpy(&ctx->buf[partial], data, len); in sha1_update() 205 size_t partial = ctx->bytecount % SHA1_BLOCK_SIZE; in __sha1_final() local 207 ctx->buf[partial++] = 0x80; in __sha1_final() 208 if (partial > SHA1_BLOCK_SIZE - 8) { in __sha1_final() [all …]
|
H A D | md5.c | 146 size_t partial = ctx->bytecount % MD5_BLOCK_SIZE; in md5_update() local 150 if (partial + len >= MD5_BLOCK_SIZE) { in md5_update() 153 if (partial) { in md5_update() 154 size_t l = MD5_BLOCK_SIZE - partial; in md5_update() 156 memcpy(&ctx->buf[partial], data, l); in md5_update() 170 partial = 0; in md5_update() 173 memcpy(&ctx->buf[partial], data, len); in md5_update() 180 size_t partial = ctx->bytecount % MD5_BLOCK_SIZE; in __md5_final() local 182 ctx->buf[partial++] = 0x80; in __md5_final() 183 if (partial > MD5_BLOCK_SIZE - 8) { in __md5_final() [all …]
|
H A D | sha512.c | 164 size_t partial = ctx->bytecount_lo % SHA512_BLOCK_SIZE; in __sha512_update() local 169 if (partial + len >= SHA512_BLOCK_SIZE) { in __sha512_update() 172 if (partial) { in __sha512_update() 173 size_t l = SHA512_BLOCK_SIZE - partial; in __sha512_update() 175 memcpy(&ctx->buf[partial], data, l); in __sha512_update() 189 partial = 0; in __sha512_update() 192 memcpy(&ctx->buf[partial], data, len); in __sha512_update() 201 size_t partial = ctx->bytecount_lo % SHA512_BLOCK_SIZE; in __sha512_final() local 203 ctx->buf[partial++] = 0x80; in __sha512_final() 204 if (partial > SHA512_BLOCK_SIZE - 16) { in __sha512_final() [all …]
|
H A D | sha256.c | 186 size_t partial = ctx->bytecount % SHA256_BLOCK_SIZE; in __sha256_update() local 190 if (partial + len >= SHA256_BLOCK_SIZE) { in __sha256_update() 193 if (partial) { in __sha256_update() 194 size_t l = SHA256_BLOCK_SIZE - partial; in __sha256_update() 196 memcpy(&ctx->buf[partial], data, l); in __sha256_update() 210 partial = 0; in __sha256_update() 213 memcpy(&ctx->buf[partial], data, len); in __sha256_update() 221 size_t partial = ctx->bytecount % SHA256_BLOCK_SIZE; in __sha256_final() local 223 ctx->buf[partial++] = 0x80; in __sha256_final() 224 if (partial > SHA256_BLOCK_SIZE - 8) { in __sha256_final() [all …]
|
H A D | chacha20poly1305.c | 221 size_t partial = 0; in chacha20poly1305_crypt_sg_inplace() local 267 if (unlikely(partial)) { in chacha20poly1305_crypt_sg_inplace() 268 size_t l = min(length, CHACHA_BLOCK_SIZE - partial); in chacha20poly1305_crypt_sg_inplace() 270 crypto_xor(addr, b.chacha_stream + partial, l); in chacha20poly1305_crypt_sg_inplace() 271 partial = (partial + l) & (CHACHA_BLOCK_SIZE - 1); in chacha20poly1305_crypt_sg_inplace() 291 partial = length; in chacha20poly1305_crypt_sg_inplace()
|
/linux/Documentation/devicetree/bindings/fpga/ |
H A D | fpga-region.yaml | 25 FPGA Regions represent FPGA's and partial reconfiguration regions of FPGA's in 41 Partial Reconfiguration (PR) 46 Partial Reconfiguration Region (PRR) 58 * Also called a "partial bit stream" 71 * During Partial Reconfiguration of a specific region, that region's bridge 86 * A base image may set up a set of partial reconfiguration regions that may 157 For partial reconfiguration (PR), each PR region will have an FPGA Region. 194 * Partial reconfiguration with bridges in the FPGA 198 region while the buses are enabled for other sections. Before any partial 207 constraints required to make partial reconfiguration work[1] [2] [3], but a few [all …]
|
/linux/fs/ext2/ |
H A D | inode.c | 325 * @partial: pointer to the last triple within a chain 331 Indirect *partial) in ext2_find_goal() argument 346 return ext2_find_near(inode, partial); in ext2_find_goal() 466 * we had read the existing part of chain and partial points to the last 632 Indirect *partial; in ext2_get_blocks() local 648 partial = ext2_get_branch(inode, depth, offsets, chain, &err); in ext2_get_blocks() 650 if (!partial) { in ext2_get_blocks() 666 partial = chain + depth - 1; in ext2_get_blocks() 696 if (err == -EAGAIN || !verify_chain(chain, partial)) { in ext2_get_blocks() 697 while (partial > chain) { in ext2_get_blocks() [all …]
|
/linux/Documentation/ABI/testing/ |
H A D | sysfs-kernel-slab | 96 allocation from a partial or new slab. It can be written to 179 The deactivate_to_head file shows how many times a partial cpu 180 slab was deactivated and added to the head of its node's partial 190 The deactivate_to_tail file shows how many times a partial cpu 191 slab was deactivated and added to the tail of its node's partial 212 partial list. It can be written to clear the current count. 255 its node's partial list. It can be written to clear the current 277 using the slow path (i.e. to a full or partial slab). It can 297 remain on a node's partial list to avoid the overhead of 326 objects are on partial slabs and from which nodes they are [all …]
|
/linux/arch/x86/include/asm/ |
H A D | unwind.h | 69 * If 'partial' returns true, only the iret frame registers are valid. 72 bool *partial) in unwind_get_entry_regs() argument 77 if (partial) { in unwind_get_entry_regs() 79 *partial = !state->full_regs; in unwind_get_entry_regs() 81 *partial = false; in unwind_get_entry_regs() 89 bool *partial) in unwind_get_entry_regs() argument
|
/linux/tools/mm/ |
H A D | slabinfo.c | 36 unsigned long partial, objects, slabs, objects_partial, objects_total; member 128 "-P|--partial Sort by number of partial slabs\n" in usage() 152 "\nSorting options (--Loss, --Size, --Partial) are mutually exclusive\n" in usage() 419 printf("%-21s ", "Partial slabs"); in slab_numa() 503 printf("Add partial %8lu %8lu %3lu %3lu\n", in slab_stats() 508 printf("Remove partial %8lu %8lu %3lu %3lu\n", in slab_stats() 513 printf("Cpu partial list %8lu %8lu %3lu %3lu\n", in slab_stats() 538 printf("Moved to head of partial list %7lu %3lu%%\n", in slab_stats() 540 printf("Moved to tail of partial list %7lu %3lu%%\n", in slab_stats() 579 s->slab_size, s->slabs - s->partial - s->cpu_slabs, in report() [all …]
|
/linux/Documentation/driver-api/md/ |
H A D | raid5-ppl.rst | 2 Partial Parity Log 5 Partial Parity Log (PPL) is a feature available for RAID5 arrays. The issue 15 Partial parity for a write operation is the XOR of stripe data chunks not 17 write hole. XORing partial parity with the modified chunks produces parity for 26 When handling a write request PPL writes partial parity before new data and
|
/linux/net/sctp/ |
H A D | ulpqueue.c | 122 /* Clear the partial delivery mode for this socket. Note: This 123 * assumes that no association is currently in partial delivery mode. 207 /* If we are in partial delivery mode, post to the lobby until in sctp_ulpq_tail_event() 208 * partial delivery is cleared, unless, of course _this_ is in sctp_ulpq_tail_event() 209 * the association the cause of the partial delivery. in sctp_ulpq_tail_event() 216 /* If the association is in partial delivery, we in sctp_ulpq_tail_event() 244 /* Did we just complete partial delivery and need to get in sctp_ulpq_tail_event() 421 * There is a potential to do partial delivery if user sets in sctp_ulpq_retrieve_reassembled() 471 /* Make sure we can enter partial deliver. in sctp_ulpq_retrieve_reassembled() 472 * We can trigger partial delivery only if framgent in sctp_ulpq_retrieve_reassembled() [all …]
|
/linux/drivers/usb/storage/ |
H A D | freecom.c | 227 unsigned int partial; in freecom_transport() local 266 FCM_STATUS_PACKET_LENGTH, &partial); in freecom_transport() 267 usb_stor_dbg(us, "foo Status result %d %u\n", result, partial); in freecom_transport() 271 US_DEBUG(pdump(us, (void *)fst, partial)); in freecom_transport() 308 FCM_STATUS_PACKET_LENGTH, &partial); in freecom_transport() 310 usb_stor_dbg(us, "bar Status result %d %u\n", result, partial); in freecom_transport() 314 US_DEBUG(pdump(us, (void *)fst, partial)); in freecom_transport() 317 if (partial != 4) in freecom_transport() 376 FCM_PACKET_LENGTH, &partial); in freecom_transport() 377 US_DEBUG(pdump(us, (void *)fst, partial)); in freecom_transport() [all …]
|
/linux/drivers/s390/crypto/ |
H A D | pkey_sysfs.c | 46 * The implementation can not deal with partial reads, because a new random 47 * protected key blob is generated with each read. In case of partial reads 103 * The implementation can not deal with partial reads, because a new random 104 * protected key blob is generated with each read. In case of partial reads 145 * The implementation can not deal with partial reads, because a new random 146 * protected key blob is generated with each read. In case of partial reads 305 * The implementation can not deal with partial reads, because a new random 306 * protected key blob is generated with each read. In case of partial reads 416 * The implementation can not deal with partial reads, because a new random 417 * secure key blob is generated with each read. In case of partial reads [all …]
|
/linux/arch/sparc/include/uapi/asm/ |
H A D | asi.h | 254 #define ASI_PST8_P 0xc0 /* Primary, 8 8-bit, partial */ 255 #define ASI_PST8_S 0xc1 /* Secondary, 8 8-bit, partial */ 256 #define ASI_PST16_P 0xc2 /* Primary, 4 16-bit, partial */ 257 #define ASI_PST16_S 0xc3 /* Secondary, 4 16-bit, partial */ 258 #define ASI_PST32_P 0xc4 /* Primary, 2 32-bit, partial */ 259 #define ASI_PST32_S 0xc5 /* Secondary, 2 32-bit, partial */ 260 #define ASI_PST8_PL 0xc8 /* Primary, 8 8-bit, partial, L */ 261 #define ASI_PST8_SL 0xc9 /* Secondary, 8 8-bit, partial, L */ 262 #define ASI_PST16_PL 0xca /* Primary, 4 16-bit, partial, L */ 263 #define ASI_PST16_SL 0xcb /* Secondary, 4 16-bit, partial, L */ [all …]
|
/linux/drivers/fpga/ |
H A D | altera-pr-ip-core.c | 3 * Driver for Altera Partial Reconfiguration IP Core 85 dev_err(&mgr->dev, "%s Partial Reconfiguration flag not set\n", in alt_pr_fpga_write_init() 94 "%s Partial Reconfiguration already started\n", in alt_pr_fpga_write_init() 156 "successful partial reconfiguration\n"); in alt_pr_fpga_write_complete() 200 MODULE_DESCRIPTION("Altera Partial Reconfiguration IP Core");
|
/linux/tools/testing/selftests/bpf/prog_tests/ |
H A D | sockmap_strp.c | 228 /* Test strparser with partial read */ 250 /* 1.1 send partial head, 1 byte header left */ in test_sockmap_strp_partial_read() 254 if (!ASSERT_EQ(-1, recvd, "partial head sent, expected no data")) in test_sockmap_strp_partial_read() 263 /* 2.1 send partial head, 1 byte header left */ in test_sockmap_strp_partial_read() 267 /* 2.2 send remaining head and partial body, 1 byte body left */ in test_sockmap_strp_partial_read() 271 if (!ASSERT_EQ(-1, recvd, "partial body sent, expected no data")) in test_sockmap_strp_partial_read() 413 /* partial read */ in test_sockmap_strp_verdict() 420 if (!ASSERT_EQ(recvd, STRP_PKT_FULL_LEN, "partial recv_timeout(c1)") || in test_sockmap_strp_verdict() 422 "partial received data does not match the sent data")) in test_sockmap_strp_verdict() 448 if (test__start_subtest("sockmap strp tcp partial read")) in test_sockmap_strp()
|
/linux/arch/alpha/lib/ |
H A D | stxcpy.S | 68 /* Take care of the final (partial) word store. 81 /* We're doing a partial word store and so need to combine 108 /* We are co-aligned; take care of a partial first word. */ 191 If it resides in the lower half, we have one (probably partial) 193 have one full and one partial word left to write out. 200 or t0, t1, t1 # e1 : first (partial) source word complete 211 /* Take care of a final (probably partial) result word. 266 /* At this point we've found a zero in the first partial word of
|
H A D | stxncpy.S | 78 /* Take care of the final (partial) word store. At this point 94 /* We're doing a partial word store and so need to combine 133 /* We are co-aligned; take care of a partial first word. */ 175 cmpbge zero, t2, t8 # e0 : find nulls in second partial 222 If it resides in the lower half, we have one (probably partial) 224 have one full and one partial word left to write out. 231 or t0, t1, t0 # e0 : first (partial) source word complete 247 /* Take care of a final (probably partial) result word. 326 /* At this point we've found a zero in the first partial word of
|
H A D | ev6-stxcpy.S | 84 /* Take care of the final (partial) word store. 96 /* We're doing a partial word store and so need to combine 128 /* We are co-aligned; take care of a partial first word. */ 219 If it resides in the lower half, we have one (probably partial) 221 have one full and one partial word left to write out. 228 or t0, t1, t1 # E : first (partial) source word complete (stall) 238 /* Take care of a final (probably partial) result word. 296 /* At this point we've found a zero in the first partial word of
|
H A D | ev6-stxncpy.S | 105 /* Take care of the final (partial) word store. At this point 120 /* We're doing a partial word store and so need to combine 166 /* We are co-aligned; take care of a partial first word. */ 217 cmpbge zero, t2, t8 # E : find nulls in second partial 266 If it resides in the lower half, we have one (probably partial) 268 have one full and one partial word left to write out. 275 or t0, t1, t0 # E : first (partial) source word complete 291 /* Take care of a final (probably partial) result word. 375 /* At this point we've found a zero in the first partial word of
|
/linux/drivers/iommu/ |
H A D | io-pgfault.c | 75 list_add(&iopf->list, &fault_param->partial); in report_partial_fault() 103 /* See if we have partial faults for this group */ in iopf_group_alloc() 105 list_for_each_entry_safe(iopf, next, &iopf_param->partial, list) { in iopf_group_alloc() 250 * leaving, otherwise partial faults will be stuck. in iommu_report_device_fault() 346 * iopf_queue_discard_partial - Remove all pending partial fault 347 * @queue: the queue whose partial faults need to be discarded 350 * lost and the IOMMU driver calls this to discard all partial faults. The 366 list_for_each_entry_safe(iopf, next, &iopf_param->partial, in iopf_queue_discard_partial() 411 INIT_LIST_HEAD(&fault_param->partial); in iopf_queue_add_device() 468 list_for_each_entry_safe(partial_iopf, next, &fault_param->partial, list) in iopf_queue_remove_device()
|
/linux/tools/testing/selftests/drivers/net/hw/ |
H A D | tso.py | 159 ethtool(f"-K {cfg.ifname} tx-gso-partial off") 162 ethtool(f"-K {cfg.ifname} tx-gso-partial on") 197 ksft_pr("partial gso feature detection may be impacted") 199 # Check which features are supported via GSO partial 201 if 'tx-gso-partial' in cfg.hw_features: 202 ethtool(f"-K {cfg.ifname} tx-gso-partial off") 209 ethtool(f"-K {cfg.ifname} tx-gso-partial on")
|