| /linux/fs/ext4/ |
| H A D | indirect.c | 252 Indirect *partial) in ext4_find_goal() argument 260 goal = ext4_find_near(inode, partial); in ext4_find_goal() 538 Indirect *partial; in ext4_ind_map_blocks() local 554 partial = ext4_get_branch(inode, depth, offsets, chain, &err); in ext4_ind_map_blocks() 557 if (!partial) { in ext4_ind_map_blocks() 586 for (i = partial - chain + 1; i < depth; i++) in ext4_ind_map_blocks() 620 ar.goal = ext4_find_goal(inode, map->m_lblk, partial); in ext4_ind_map_blocks() 623 indirect_blks = (chain + depth) - partial - 1; in ext4_ind_map_blocks() 629 ar.len = ext4_blks_to_allocate(partial, indirect_blks, in ext4_ind_map_blocks() 636 offsets + (partial - chain), partial); in ext4_ind_map_blocks() [all …]
|
| H A D | extents.c | 2465 struct partial_cluster *partial, in ext4_remove_blocks() argument 2496 trace_ext4_remove_blocks(inode, ex, from, to, partial); in ext4_remove_blocks() 2504 if (partial->state != initial && in ext4_remove_blocks() 2505 partial->pclu != EXT4_B2C(sbi, last_pblk)) { in ext4_remove_blocks() 2506 if (partial->state == tofree) { in ext4_remove_blocks() 2508 if (ext4_is_pending(inode, partial->lblk)) in ext4_remove_blocks() 2511 EXT4_C2B(sbi, partial->pclu), in ext4_remove_blocks() 2514 ext4_rereserve_cluster(inode, partial->lblk); in ext4_remove_blocks() 2516 partial->state = initial; in ext4_remove_blocks() 2533 (partial->state != nofree)) { in ext4_remove_blocks() [all …]
|
| /linux/fs/minix/ |
| H A D | itree_common.c | 158 Indirect *partial; in get_block() local 166 partial = get_branch(inode, depth, offsets, chain, &err); in get_block() 169 if (!partial) { in get_block() 173 partial = chain+depth-1; /* the whole chain */ in get_block() 180 while (partial > chain) { in get_block() 181 brelse(partial->bh); in get_block() 182 partial--; in get_block() 196 left = (chain + depth) - partial; in get_block() 197 err = alloc_branch(inode, left, offsets+(partial-chain), partial); in get_block() 201 if (splice_branch(inode, chain, partial, left) < 0) in get_block() [all …]
|
| /linux/fs/ext2/ |
| H A D | inode.c | 331 Indirect *partial) in ext2_find_goal() argument 346 return ext2_find_near(inode, partial); in ext2_find_goal() 632 Indirect *partial; in ext2_get_blocks() local 648 partial = ext2_get_branch(inode, depth, offsets, chain, &err); in ext2_get_blocks() 650 if (!partial) { in ext2_get_blocks() 666 partial = chain + depth - 1; in ext2_get_blocks() 696 if (err == -EAGAIN || !verify_chain(chain, partial)) { in ext2_get_blocks() 697 while (partial > chain) { in ext2_get_blocks() 698 brelse(partial->bh); in ext2_get_blocks() 699 partial--; in ext2_get_blocks() [all …]
|
| /linux/lib/crypto/ |
| H A D | sha1.c | 143 size_t partial = ctx->bytecount % SHA1_BLOCK_SIZE; 147 if (partial + len >= SHA1_BLOCK_SIZE) { in sha1_blocks_generic() 150 if (partial) { in sha1_blocks_generic() 151 size_t l = SHA1_BLOCK_SIZE - partial; in sha1_blocks_generic() 153 memcpy(&ctx->buf[partial], data, l); in sha1_blocks_generic() 167 partial = 0; in sha1_init() 170 memcpy(&ctx->buf[partial], data, len); in sha1_update() 177 size_t partial = ctx->bytecount % SHA1_BLOCK_SIZE; in sha1_update() 179 ctx->buf[partial++] = 0x80; in sha1_update() 180 if (partial > SHA1_BLOCK_SIZ in sha1_update() 172 size_t partial = ctx->bytecount % SHA1_BLOCK_SIZE; sha1_update() local 206 size_t partial = ctx->bytecount % SHA1_BLOCK_SIZE; __sha1_final() local [all...] |
| H A D | md5.c | 146 size_t partial = ctx->bytecount % MD5_BLOCK_SIZE; in md5_update() local 150 if (partial + len >= MD5_BLOCK_SIZE) { in md5_update() 153 if (partial) { in md5_update() 154 size_t l = MD5_BLOCK_SIZE - partial; in md5_update() 156 memcpy(&ctx->buf[partial], data, l); in md5_update() 170 partial = 0; in md5_update() 173 memcpy(&ctx->buf[partial], data, len); in md5_update() 180 size_t partial = ctx->bytecount % MD5_BLOCK_SIZE; in __md5_final() local 182 ctx->buf[partial++] = 0x80; in __md5_final() 183 if (partial > MD5_BLOCK_SIZ in __md5_final() [all...] |
| H A D | sha512.c | 165 size_t partial = ctx->bytecount_lo % SHA512_BLOCK_SIZE; in __sha512_update() local 170 if (partial + len >= SHA512_BLOCK_SIZE) { in __sha512_update() 173 if (partial) { in __sha512_update() 174 size_t l = SHA512_BLOCK_SIZE - partial; in __sha512_update() 176 memcpy(&ctx->buf[partial], data, l); in __sha512_update() 190 partial = 0; in __sha512_update() 193 memcpy(&ctx->buf[partial], data, len); in __sha512_update() 202 size_t partial = ctx->bytecount_lo % SHA512_BLOCK_SIZE; in __sha512_final() local 204 ctx->buf[partial++] = 0x80; in __sha512_final() 205 if (partial > SHA512_BLOCK_SIZE - 16) { in __sha512_final() [all …]
|
| H A D | sha256.c | 187 size_t partial = ctx->bytecount % SHA256_BLOCK_SIZE; in __sha256_update() local 191 if (partial + len >= SHA256_BLOCK_SIZE) { in __sha256_update() 194 if (partial) { in __sha256_update() 195 size_t l = SHA256_BLOCK_SIZE - partial; in __sha256_update() 197 memcpy(&ctx->buf[partial], data, l); in __sha256_update() 211 partial = 0; in __sha256_update() 214 memcpy(&ctx->buf[partial], data, len); in __sha256_update() 222 size_t partial = ctx->bytecount % SHA256_BLOCK_SIZE; in __sha256_final() local 224 ctx->buf[partial++] = 0x80; in __sha256_final() 225 if (partial > SHA256_BLOCK_SIZE - 8) { in __sha256_final() [all …]
|
| H A D | polyval.c | 260 if (unlikely(ctx->partial)) { in polyval_update() 261 size_t n = min(len, POLYVAL_BLOCK_SIZE - ctx->partial); in polyval_update() 265 ctx->acc.bytes[ctx->partial++] ^= *data++; in polyval_update() 266 if (ctx->partial < POLYVAL_BLOCK_SIZE) in polyval_update() 279 ctx->partial = len; in polyval_update() 285 if (unlikely(ctx->partial)) in polyval_final()
|
| H A D | chacha20poly1305.c | 221 size_t partial = 0; in chacha20poly1305_crypt_sg_inplace() local 267 if (unlikely(partial)) { in chacha20poly1305_crypt_sg_inplace() 268 size_t l = min(length, CHACHA_BLOCK_SIZE - partial); in chacha20poly1305_crypt_sg_inplace() 270 crypto_xor(addr, b.chacha_stream + partial, l); in chacha20poly1305_crypt_sg_inplace() 271 partial = (partial + l) & (CHACHA_BLOCK_SIZE - 1); in chacha20poly1305_crypt_sg_inplace() 291 partial = length; in chacha20poly1305_crypt_sg_inplace()
|
| /linux/drivers/usb/storage/ |
| H A D | freecom.c | 227 unsigned int partial; in freecom_transport() local 266 FCM_STATUS_PACKET_LENGTH, &partial); in freecom_transport() 267 usb_stor_dbg(us, "foo Status result %d %u\n", result, partial); in freecom_transport() 271 US_DEBUG(pdump(us, (void *)fst, partial)); in freecom_transport() 308 FCM_STATUS_PACKET_LENGTH, &partial); in freecom_transport() 310 usb_stor_dbg(us, "bar Status result %d %u\n", result, partial); in freecom_transport() 314 US_DEBUG(pdump(us, (void *)fst, partial)); in freecom_transport() 317 if (partial != 4) in freecom_transport() 376 FCM_PACKET_LENGTH, &partial); in freecom_transport() 377 US_DEBUG(pdump(us, (void *)fst, partial)); in freecom_transport() [all …]
|
| H A D | transport.c | 257 unsigned int length, int result, unsigned int partial) in interpret_urb_result() argument 260 result, partial, length); in interpret_urb_result() 265 if (partial != length) { in interpret_urb_result() 469 unsigned int partial; in usb_stor_bulk_srb() local 472 &partial); in usb_stor_bulk_srb() 474 scsi_set_resid(srb, scsi_bufflen(srb) - partial); in usb_stor_bulk_srb() 492 unsigned int partial; in usb_stor_bulk_transfer_sg() local 499 length_left, &partial); in usb_stor_bulk_transfer_sg() 500 length_left -= partial; in usb_stor_bulk_transfer_sg() 504 length_left, &partial); in usb_stor_bulk_transfer_sg() [all …]
|
| /linux/net/smc/ |
| H A D | smc_rx.c | 154 struct partial_page *partial; in smc_rx_splice() local 167 partial = kzalloc_objs(*partial, nr_pages); in smc_rx_splice() 168 if (!partial) in smc_rx_splice() 184 partial[0].offset = src - (char *)smc->conn.rmb_desc->cpu_addr; in smc_rx_splice() 185 partial[0].len = len; in smc_rx_splice() 186 partial[0].private = (unsigned long)priv[0]; in smc_rx_splice() 197 partial[i].offset = offset; in smc_rx_splice() 198 partial[i].len = size; in smc_rx_splice() 199 partial[ in smc_rx_splice() [all...] |
| /linux/arch/x86/include/asm/ |
| H A D | unwind.h | 72 bool *partial) in unwind_get_entry_regs() argument 77 if (partial) { in unwind_get_entry_regs() 79 *partial = !state->full_regs; in unwind_get_entry_regs() 81 *partial = false; in unwind_get_entry_regs() 89 bool *partial) in unwind_get_entry_regs() argument
|
| /linux/tools/mm/ |
| H A D | slabinfo.c | 36 unsigned long partial, objects, slabs, objects_partial, objects_total; member 128 "-P|--partial Sort by number of partial slabs\n" in usage() 503 printf("Add partial %8lu %8lu %3lu %3lu\n", in slab_stats() 508 printf("Remove partial %8lu %8lu %3lu %3lu\n", in slab_stats() 513 printf("Cpu partial list %8lu %8lu %3lu %3lu\n", in slab_stats() 538 printf("Moved to head of partial list %7lu %3lu%%\n", in slab_stats() 540 printf("Moved to tail of partial list %7lu %3lu%%\n", in slab_stats() 579 s->slab_size, s->slabs - s->partial - s->cpu_slabs, in report() 582 page_size << s->order, s->partial, onof in report() 1358 _xtotals(char * heading,char * underline,int loss,int size,int partial) _xtotals() argument [all...] |
| /linux/Documentation/ABI/testing/ |
| H A D | sysfs-kernel-slab | 96 allocation from a partial or new slab. It can be written to 179 The deactivate_to_head file shows how many times a partial cpu 180 slab was deactivated and added to the head of its node's partial 190 The deactivate_to_tail file shows how many times a partial cpu 191 slab was deactivated and added to the tail of its node's partial 212 partial list. It can be written to clear the current count. 255 its node's partial list. It can be written to clear the current 277 using the slow path (i.e. to a full or partial slab). It can 297 remain on a node's partial list to avoid the overhead of 326 objects are on partial slabs and from which nodes they are [all …]
|
| /linux/tools/testing/selftests/syscall_user_dispatch/ |
| H A D | sud_benchmark.c | 97 double partial = 0; in perf_syscall() local 100 partial += one_sysinfo_step()/(CALIBRATION_STEP*factor); in perf_syscall() 101 return partial; in perf_syscall()
|
| /linux/crypto/ |
| H A D | md5.c | 28 unsigned int partial; in __crypto_md5_export() local 31 partial = ctx.bytecount % MD5_BLOCK_SIZE; in __crypto_md5_export() 32 ctx.bytecount -= partial; in __crypto_md5_export() 35 *p = partial; in __crypto_md5_export()
|
| H A D | sha1.c | 31 unsigned int partial; in __crypto_sha1_export() local 34 partial = ctx.bytecount % SHA1_BLOCK_SIZE; in __crypto_sha1_export() 35 ctx.bytecount -= partial; in __crypto_sha1_export() 38 *p = partial; in __crypto_sha1_export()
|
| H A D | sha256.c | 32 unsigned int partial; in __crypto_sha256_export() local 35 partial = ctx.bytecount % SHA256_BLOCK_SIZE; in __crypto_sha256_export() 36 ctx.bytecount -= partial; in __crypto_sha256_export() 39 *p = partial; in __crypto_sha256_export()
|
| H A D | sha512.c | 32 unsigned int partial; in __crypto_sha512_export() local 35 partial = ctx.bytecount_lo % SHA512_BLOCK_SIZE; in __crypto_sha512_export() 36 ctx.bytecount_lo -= partial; in __crypto_sha512_export() 39 *p = partial; in __crypto_sha512_export()
|
| /linux/drivers/iommu/ |
| H A D | io-pgfault.c | 75 list_add(&iopf->list, &fault_param->partial); in report_partial_fault() 103 /* See if we have partial faults for this group */ in iopf_group_alloc() 105 list_for_each_entry_safe(iopf, next, &iopf_param->partial, list) { in iopf_group_alloc() 250 * leaving, otherwise partial faults will be stuck. in iommu_report_device_fault() 346 * iopf_queue_discard_partial - Remove all pending partial fault 347 * @queue: the queue whose partial faults need to be discarded 350 * lost and the IOMMU driver calls this to discard all partial faults. The 366 list_for_each_entry_safe(iopf, next, &iopf_param->partial, in iopf_queue_discard_partial() 411 INIT_LIST_HEAD(&fault_param->partial); in iopf_queue_add_device() 468 list_for_each_entry_safe(partial_iopf, next, &fault_param->partial, lis in iopf_queue_remove_device() [all...] |
| /linux/drivers/comedi/drivers/ni_routing/tools/ |
| H A D | Makefile | 69 clean-partial : 73 clean : clean-partial
|
| /linux/include/net/ |
| H A D | gso.h | 69 __wsum partial = SKB_GSO_CB(skb)->csum; in gso_make_checksum() local 74 return csum_fold(csum_partial(csum_start, plen, partial)); in gso_make_checksum()
|
| /linux/lib/tests/ |
| H A D | fortify_kunit.c | 711 int i, partial; in fortify_test_strncat() local 714 partial = sizeof(src) / 2 - 1; in fortify_test_strncat() 715 for (i = 0; i < partial; i++) in fortify_test_strncat() 726 KUNIT_ASSERT_TRUE(test, strncat(pad.buf, src, partial) == pad.buf); in fortify_test_strncat() 730 KUNIT_ASSERT_TRUE(test, strncat(pad.buf, src, partial) == pad.buf); in fortify_test_strncat() 784 int i, partial; in fortify_test_strlcat() local 790 partial = sizeof(src) / 2 - 1; in fortify_test_strlcat() 791 for (i = 0; i < partial; i++) in fortify_test_strlcat() 802 KUNIT_ASSERT_EQ(test, strlcat(pad.buf, src, len), partial); in fortify_test_strlcat() 806 KUNIT_ASSERT_EQ(test, strlcat(pad.buf, src, len), partial * 2); in fortify_test_strlcat() [all …]
|