| /linux/fs/ext4/ |
| H A D | indirect.c | 252 Indirect *partial) in ext4_find_goal() argument 260 goal = ext4_find_near(inode, partial); in ext4_find_goal() 538 Indirect *partial; in ext4_ind_map_blocks() local 554 partial = ext4_get_branch(inode, depth, offsets, chain, &err); in ext4_ind_map_blocks() 557 if (!partial) { in ext4_ind_map_blocks() 586 for (i = partial - chain + 1; i < depth; i++) in ext4_ind_map_blocks() 620 ar.goal = ext4_find_goal(inode, map->m_lblk, partial); in ext4_ind_map_blocks() 623 indirect_blks = (chain + depth) - partial - 1; in ext4_ind_map_blocks() 629 ar.len = ext4_blks_to_allocate(partial, indirect_blks, in ext4_ind_map_blocks() 636 offsets + (partial - chain), partial); in ext4_ind_map_blocks() [all …]
|
| /linux/fs/minix/ |
| H A D | itree_common.c | 158 Indirect *partial; in get_block() local 166 partial = get_branch(inode, depth, offsets, chain, &err); in get_block() 169 if (!partial) { in get_block() 173 partial = chain+depth-1; /* the whole chain */ in get_block() 180 while (partial > chain) { in get_block() 181 brelse(partial->bh); in get_block() 182 partial--; in get_block() 196 left = (chain + depth) - partial; in get_block() 197 err = alloc_branch(inode, left, offsets+(partial-chain), partial); in get_block() 201 if (splice_branch(inode, chain, partial, left) < 0) in get_block() [all …]
|
| /linux/fs/ext2/ |
| H A D | inode.c | 331 Indirect *partial) in ext2_find_goal() argument 346 return ext2_find_near(inode, partial); in ext2_find_goal() 632 Indirect *partial; in ext2_get_blocks() local 648 partial = ext2_get_branch(inode, depth, offsets, chain, &err); in ext2_get_blocks() 650 if (!partial) { in ext2_get_blocks() 666 partial = chain + depth - 1; in ext2_get_blocks() 696 if (err == -EAGAIN || !verify_chain(chain, partial)) { in ext2_get_blocks() 697 while (partial > chain) { in ext2_get_blocks() 698 brelse(partial->bh); in ext2_get_blocks() 699 partial--; in ext2_get_blocks() [all …]
|
| /linux/lib/crypto/ |
| H A D | sha1.c | 172 size_t partial = ctx->bytecount % SHA1_BLOCK_SIZE; in sha1_update() local 176 if (partial + len >= SHA1_BLOCK_SIZE) { in sha1_update() 179 if (partial) { in sha1_update() 180 size_t l = SHA1_BLOCK_SIZE - partial; in sha1_update() 182 memcpy(&ctx->buf[partial], data, l); in sha1_update() 196 partial = 0; in sha1_update() 199 memcpy(&ctx->buf[partial], data, len); in sha1_update() 206 size_t partial = ctx->bytecount % SHA1_BLOCK_SIZE; in __sha1_final() local 208 ctx->buf[partial++] = 0x80; in __sha1_final() 209 if (partial > SHA1_BLOCK_SIZE - 8) { in __sha1_final() [all …]
|
| H A D | md5.c | 146 size_t partial = ctx->bytecount % MD5_BLOCK_SIZE; in md5_update() local 150 if (partial + len >= MD5_BLOCK_SIZE) { in md5_update() 153 if (partial) { in md5_update() 154 size_t l = MD5_BLOCK_SIZE - partial; in md5_update() 156 memcpy(&ctx->buf[partial], data, l); in md5_update() 170 partial = 0; in md5_update() 173 memcpy(&ctx->buf[partial], data, len); in md5_update() 180 size_t partial = ctx->bytecount % MD5_BLOCK_SIZE; in __md5_final() local 182 ctx->buf[partial++] = 0x80; in __md5_final() 183 if (partial > MD5_BLOCK_SIZE - 8) { in __md5_final() [all …]
|
| H A D | sha512.c | 165 size_t partial = ctx->bytecount_lo % SHA512_BLOCK_SIZE; in __sha512_update() local 170 if (partial + len >= SHA512_BLOCK_SIZE) { in __sha512_update() 173 if (partial) { in __sha512_update() 174 size_t l = SHA512_BLOCK_SIZE - partial; in __sha512_update() 176 memcpy(&ctx->buf[partial], data, l); in __sha512_update() 190 partial = 0; in __sha512_update() 193 memcpy(&ctx->buf[partial], data, len); in __sha512_update() 202 size_t partial = ctx->bytecount_lo % SHA512_BLOCK_SIZE; in __sha512_final() local 204 ctx->buf[partial++] = 0x80; in __sha512_final() 205 if (partial > SHA512_BLOCK_SIZE - 16) { in __sha512_final() [all …]
|
| H A D | sha256.c | 187 size_t partial = ctx->bytecount % SHA256_BLOCK_SIZE; in __sha256_update() local 191 if (partial + len >= SHA256_BLOCK_SIZE) { in __sha256_update() 194 if (partial) { in __sha256_update() 195 size_t l = SHA256_BLOCK_SIZE - partial; in __sha256_update() 197 memcpy(&ctx->buf[partial], data, l); in __sha256_update() 211 partial = 0; in __sha256_update() 214 memcpy(&ctx->buf[partial], data, len); in __sha256_update() 222 size_t partial = ctx->bytecount % SHA256_BLOCK_SIZE; in __sha256_final() local 224 ctx->buf[partial++] = 0x80; in __sha256_final() 225 if (partial > SHA256_BLOCK_SIZE - 8) { in __sha256_final() [all …]
|
| H A D | polyval.c | 260 if (unlikely(ctx->partial)) { in polyval_update() 261 size_t n = min(len, POLYVAL_BLOCK_SIZE - ctx->partial); in polyval_update() 265 ctx->acc.bytes[ctx->partial++] ^= *data++; in polyval_update() 266 if (ctx->partial < POLYVAL_BLOCK_SIZE) in polyval_update() 279 ctx->partial = len; in polyval_update() 285 if (unlikely(ctx->partial)) in polyval_final()
|
| H A D | chacha20poly1305.c | 221 size_t partial = 0; in chacha20poly1305_crypt_sg_inplace() local 267 if (unlikely(partial)) { in chacha20poly1305_crypt_sg_inplace() 268 size_t l = min(length, CHACHA_BLOCK_SIZE - partial); in chacha20poly1305_crypt_sg_inplace() 270 crypto_xor(addr, b.chacha_stream + partial, l); in chacha20poly1305_crypt_sg_inplace() 271 partial = (partial + l) & (CHACHA_BLOCK_SIZE - 1); in chacha20poly1305_crypt_sg_inplace() 291 partial = length; in chacha20poly1305_crypt_sg_inplace()
|
| /linux/drivers/usb/storage/ |
| H A D | freecom.c | 227 unsigned int partial; in freecom_transport() local 266 FCM_STATUS_PACKET_LENGTH, &partial); in freecom_transport() 267 usb_stor_dbg(us, "foo Status result %d %u\n", result, partial); in freecom_transport() 271 US_DEBUG(pdump(us, (void *)fst, partial)); in freecom_transport() 308 FCM_STATUS_PACKET_LENGTH, &partial); in freecom_transport() 310 usb_stor_dbg(us, "bar Status result %d %u\n", result, partial); in freecom_transport() 314 US_DEBUG(pdump(us, (void *)fst, partial)); in freecom_transport() 317 if (partial != 4) in freecom_transport() 376 FCM_PACKET_LENGTH, &partial); in freecom_transport() 377 US_DEBUG(pdump(us, (void *)fst, partial)); in freecom_transport() [all …]
|
| H A D | transport.c | 257 unsigned int length, int result, unsigned int partial) in interpret_urb_result() argument 260 result, partial, length); in interpret_urb_result() 265 if (partial != length) { in interpret_urb_result() 469 unsigned int partial; in usb_stor_bulk_srb() local 472 &partial); in usb_stor_bulk_srb() 474 scsi_set_resid(srb, scsi_bufflen(srb) - partial); in usb_stor_bulk_srb() 492 unsigned int partial; in usb_stor_bulk_transfer_sg() local 499 length_left, &partial); in usb_stor_bulk_transfer_sg() 500 length_left -= partial; in usb_stor_bulk_transfer_sg() 504 length_left, &partial); in usb_stor_bulk_transfer_sg() [all …]
|
| /linux/arch/x86/include/asm/ |
| H A D | unwind.h | 72 bool *partial) in unwind_get_entry_regs() argument 77 if (partial) { in unwind_get_entry_regs() 79 *partial = !state->full_regs; in unwind_get_entry_regs() 81 *partial = false; in unwind_get_entry_regs() 89 bool *partial) in unwind_get_entry_regs() argument
|
| /linux/tools/mm/ |
| H A D | slabinfo.c | 36 unsigned long partial, objects, slabs, objects_partial, objects_total; member 579 s->slab_size, s->slabs - s->partial - s->cpu_slabs, in report() 582 page_size << s->order, s->partial, onoff(s->poison), in report() 627 s->partial, s->cpu_slabs); in slabcache() 670 s->slabs ? (s->partial * 100) / s->slabs : 100, in slabcache() 878 percentage_partial_slabs = s->partial * 100 / s->slabs; in totals() 890 if (s->partial < min_partial) in totals() 891 min_partial = s->partial; in totals() 915 if (s->partial > max_partial) in totals() 916 max_partial = s->partial; in totals() [all …]
|
| /linux/Documentation/ABI/testing/ |
| H A D | sysfs-kernel-slab | 96 allocation from a partial or new slab. It can be written to 179 The deactivate_to_head file shows how many times a partial cpu 180 slab was deactivated and added to the head of its node's partial 190 The deactivate_to_tail file shows how many times a partial cpu 191 slab was deactivated and added to the tail of its node's partial 212 partial list. It can be written to clear the current count. 255 its node's partial list. It can be written to clear the current 277 using the slow path (i.e. to a full or partial slab). It can 297 remain on a node's partial list to avoid the overhead of 326 objects are on partial slabs and from which nodes they are [all …]
|
| /linux/tools/testing/selftests/syscall_user_dispatch/ |
| H A D | sud_benchmark.c | 97 double partial = 0; in perf_syscall() local 100 partial += one_sysinfo_step()/(CALIBRATION_STEP*factor); in perf_syscall() 101 return partial; in perf_syscall()
|
| /linux/crypto/ |
| H A D | md5.c | 28 unsigned int partial; in __crypto_md5_export() local 31 partial = ctx.bytecount % MD5_BLOCK_SIZE; in __crypto_md5_export() 32 ctx.bytecount -= partial; in __crypto_md5_export() 35 *p = partial; in __crypto_md5_export()
|
| H A D | sha1.c | 31 unsigned int partial; in __crypto_sha1_export() local 34 partial = ctx.bytecount % SHA1_BLOCK_SIZE; in __crypto_sha1_export() 35 ctx.bytecount -= partial; in __crypto_sha1_export() 38 *p = partial; in __crypto_sha1_export()
|
| H A D | sha256.c | 32 unsigned int partial; in __crypto_sha256_export() local 35 partial = ctx.bytecount % SHA256_BLOCK_SIZE; in __crypto_sha256_export() 36 ctx.bytecount -= partial; in __crypto_sha256_export() 39 *p = partial; in __crypto_sha256_export()
|
| H A D | sha512.c | 32 unsigned int partial; in __crypto_sha512_export() local 35 partial = ctx.bytecount_lo % SHA512_BLOCK_SIZE; in __crypto_sha512_export() 36 ctx.bytecount_lo -= partial; in __crypto_sha512_export() 39 *p = partial; in __crypto_sha512_export()
|
| /linux/drivers/iommu/ |
| H A D | io-pgfault.c | 75 list_add(&iopf->list, &fault_param->partial); in report_partial_fault() 105 list_for_each_entry_safe(iopf, next, &iopf_param->partial, list) { in iopf_group_alloc() 366 list_for_each_entry_safe(iopf, next, &iopf_param->partial, in iopf_queue_discard_partial() 411 INIT_LIST_HEAD(&fault_param->partial); in iopf_queue_add_device() 468 list_for_each_entry_safe(partial_iopf, next, &fault_param->partial, list) in iopf_queue_remove_device()
|
| /linux/drivers/comedi/drivers/ni_routing/tools/ |
| H A D | Makefile | 69 clean-partial : 73 clean : clean-partial
|
| /linux/include/net/ |
| H A D | gso.h | 69 __wsum partial = SKB_GSO_CB(skb)->csum; in gso_make_checksum() local 74 return csum_fold(csum_partial(csum_start, plen, partial)); in gso_make_checksum()
|
| /linux/net/netfilter/ |
| H A D | xt_dccp.c | 56 goto partial; in dccp_find_option() 74 partial: in dccp_find_option()
|
| /linux/lib/tests/ |
| H A D | fortify_kunit.c | 711 int i, partial; in fortify_test_strncat() local 714 partial = sizeof(src) / 2 - 1; in fortify_test_strncat() 715 for (i = 0; i < partial; i++) in fortify_test_strncat() 726 KUNIT_ASSERT_TRUE(test, strncat(pad.buf, src, partial) == pad.buf); in fortify_test_strncat() 730 KUNIT_ASSERT_TRUE(test, strncat(pad.buf, src, partial) == pad.buf); in fortify_test_strncat() 784 int i, partial; in fortify_test_strlcat() local 790 partial = sizeof(src) / 2 - 1; in fortify_test_strlcat() 791 for (i = 0; i < partial; i++) in fortify_test_strlcat() 802 KUNIT_ASSERT_EQ(test, strlcat(pad.buf, src, len), partial); in fortify_test_strlcat() 806 KUNIT_ASSERT_EQ(test, strlcat(pad.buf, src, len), partial * 2); in fortify_test_strlcat() [all …]
|
| /linux/Documentation/ABI/stable/ |
| H A D | sysfs-driver-ib_srp | 43 * allow_ext_sg, whether ib_srp is allowed to include a partial 45 list. If a partial memory descriptor list has been included 51 implementations support partial memory descriptor lists the 56 only safe with partial memory descriptor list support enabled 90 Description: Whether ib_srp is allowed to include a partial memory
|