| /linux/fs/squashfs/ |
| H A D | xattr_id.c | 41 start_block = le64_to_cpu(msblk->xattr_id_table[block]); in squashfs_xattr_lookup() 48 *xattr = le64_to_cpu(id.xattr); in squashfs_xattr_lookup() 72 *xattr_table_start = le64_to_cpu(id_table->xattr_table_start); in squashfs_read_xattr_id_table() 109 start = le64_to_cpu(table[n]); in squashfs_read_xattr_id_table() 110 end = le64_to_cpu(table[n + 1]); in squashfs_read_xattr_id_table() 119 start = le64_to_cpu(table[indexes - 1]); in squashfs_read_xattr_id_table() 126 if (*xattr_table_start >= le64_to_cpu(table[0])) { in squashfs_read_xattr_id_table()
|
| H A D | export.c | 53 start = le64_to_cpu(msblk->inode_lookup_table[blk]); in squashfs_inode_lookup() 60 (u64) le64_to_cpu(ino)); in squashfs_inode_lookup() 62 return le64_to_cpu(ino); in squashfs_inode_lookup() 152 start = le64_to_cpu(table[n]); in squashfs_read_inode_lookup_table() 153 end = le64_to_cpu(table[n + 1]); in squashfs_read_inode_lookup_table() 163 start = le64_to_cpu(table[indexes - 1]); in squashfs_read_inode_lookup_table()
|
| H A D | id.c | 45 start_block = le64_to_cpu(msblk->id_table[block]); in squashfs_get_id() 97 start = le64_to_cpu(table[n]); in squashfs_read_id_index_table() 98 end = le64_to_cpu(table[n + 1]); in squashfs_read_id_index_table() 107 start = le64_to_cpu(table[indexes - 1]); in squashfs_read_id_index_table()
|
| H A D | super.c | 258 msblk->bytes_used = le64_to_cpu(sblk->bytes_used); in squashfs_fill_super() 288 root_inode = le64_to_cpu(sblk->root_inode); in squashfs_fill_super() 292 msblk->inode_table = le64_to_cpu(sblk->inode_table_start); in squashfs_fill_super() 293 msblk->directory_table = le64_to_cpu(sblk->directory_table_start); in squashfs_fill_super() 312 (u64) le64_to_cpu(sblk->fragment_table_start)); in squashfs_fill_super() 314 (u64) le64_to_cpu(sblk->id_table_start)); in squashfs_fill_super() 362 xattr_id_table_start = le64_to_cpu(sblk->xattr_id_table_start); in squashfs_fill_super() 383 le64_to_cpu(sblk->id_table_start), next_table, msblk->ids); in squashfs_fill_super() 390 next_table = le64_to_cpu(msblk->id_table[0]); in squashfs_fill_super() 393 lookup_table_start = le64_to_cpu(sblk->lookup_table_start); in squashfs_fill_super() [all …]
|
| H A D | fragment.c | 48 start_block = le64_to_cpu(msblk->fragment_index[block]); in squashfs_frag_lookup() 55 *fragment_block = le64_to_cpu(fragment_entry.start_block); in squashfs_frag_lookup() 83 if (!IS_ERR(table) && le64_to_cpu(table[0]) >= fragment_table_start) { in squashfs_read_fragment_index_table()
|
| /linux/include/linux/ |
| H A D | siphash.h | 61 return siphash_1u64(le64_to_cpu(data[0]), key); in ___siphash_aligned() 63 return siphash_2u64(le64_to_cpu(data[0]), le64_to_cpu(data[1]), in ___siphash_aligned() 66 return siphash_3u64(le64_to_cpu(data[0]), le64_to_cpu(data[1]), in ___siphash_aligned() 67 le64_to_cpu(data[2]), key); in ___siphash_aligned() 69 return siphash_4u64(le64_to_cpu(data[0]), le64_to_cpu(data[1]), in ___siphash_aligned() 70 le64_to_cpu(data[2]), le64_to_cpu(data[3]), in ___siphash_aligned()
|
| /linux/fs/smb/client/ |
| H A D | smb2maperror.c | 59 le64_to_cpu(shdr->SessionId), in map_smb2_to_linux_error() 61 le64_to_cpu(shdr->MessageId)); in map_smb2_to_linux_error() 85 le64_to_cpu(shdr->SessionId), in map_smb2_to_linux_error() 87 le64_to_cpu(shdr->MessageId), in map_smb2_to_linux_error()
|
| /linux/fs/ocfs2/ |
| H A D | inode.c | 209 if (S_ISLNK(le16_to_cpu(di->i_mode)) && le64_to_cpu(di->i_size) && in ocfs2_dinode_has_extents() 370 inode->i_rdev = huge_decode_dev(le64_to_cpu(fe->id1.dev1.i_rdev)); in ocfs2_populate_inode() 383 inode_set_atime(inode, le64_to_cpu(fe->i_atime), in ocfs2_populate_inode() 385 inode_set_mtime(inode, le64_to_cpu(fe->i_mtime), in ocfs2_populate_inode() 387 inode_set_ctime(inode, le64_to_cpu(fe->i_ctime), in ocfs2_populate_inode() 390 if (OCFS2_I(inode)->ip_blkno != le64_to_cpu(fe->i_blkno)) in ocfs2_populate_inode() 394 (unsigned long long)le64_to_cpu(fe->i_blkno)); in ocfs2_populate_inode() 424 i_size_write(inode, le64_to_cpu(fe->i_size)); in ocfs2_populate_inode() 432 i_size_write(inode, le64_to_cpu(fe->i_size)); in ocfs2_populate_inode() 438 i_size_write(inode, le64_to_cpu(fe->i_size)); in ocfs2_populate_inode() [all …]
|
| /linux/fs/quota/ |
| H A D | quota_v2.c | 237 m->dqb_itime = le64_to_cpu(d->dqb_itime); in v2r0_disk2memdqb() 240 m->dqb_curspace = le64_to_cpu(d->dqb_curspace); in v2r0_disk2memdqb() 241 m->dqb_btime = le64_to_cpu(d->dqb_btime); in v2r0_disk2memdqb() 287 m->dqb_ihardlimit = le64_to_cpu(d->dqb_ihardlimit); in v2r1_disk2memdqb() 288 m->dqb_isoftlimit = le64_to_cpu(d->dqb_isoftlimit); in v2r1_disk2memdqb() 289 m->dqb_curinodes = le64_to_cpu(d->dqb_curinodes); in v2r1_disk2memdqb() 290 m->dqb_itime = le64_to_cpu(d->dqb_itime); in v2r1_disk2memdqb() 291 m->dqb_bhardlimit = v2_qbtos(le64_to_cpu(d->dqb_bhardlimit)); in v2r1_disk2memdqb() 292 m->dqb_bsoftlimit = v2_qbtos(le64_to_cpu(d->dqb_bsoftlimit)); in v2r1_disk2memdqb() 293 m->dqb_curspace = le64_to_cpu(d->dqb_curspace); in v2r1_disk2memdqb() [all …]
|
| /linux/drivers/bus/mhi/host/ |
| H A D | debugfs.c | 69 seq_printf(m, " base: 0x%0llx len: 0x%llx", le64_to_cpu(er_ctxt->rbase), in mhi_debugfs_events_show() 70 le64_to_cpu(er_ctxt->rlen)); in mhi_debugfs_events_show() 72 seq_printf(m, " rp: 0x%llx wp: 0x%llx", le64_to_cpu(er_ctxt->rp), in mhi_debugfs_events_show() 73 le64_to_cpu(er_ctxt->wp)); in mhi_debugfs_events_show() 120 le64_to_cpu(chan_ctxt->rbase), le64_to_cpu(chan_ctxt->rlen), in mhi_debugfs_channels_show() 121 le64_to_cpu(chan_ctxt->rp), le64_to_cpu(chan_ctxt->wp)); in mhi_debugfs_channels_show()
|
| /linux/drivers/net/ethernet/apm/xgene-v2/ |
| H A D | main.c | 94 addr_hi = GET_BITS(NEXT_DESC_ADDRH, le64_to_cpu(raw_desc->m1)); in xge_refill_buffers() 95 addr_lo = GET_BITS(NEXT_DESC_ADDRL, le64_to_cpu(raw_desc->m1)); in xge_refill_buffers() 163 if (GET_BITS(E, le64_to_cpu(raw_desc->m0)) && in is_tx_slot_available() 164 (GET_BITS(PKT_SIZE, le64_to_cpu(raw_desc->m0)) == SLOT_EMPTY)) in is_tx_slot_available() 201 addr_hi = GET_BITS(NEXT_DESC_ADDRH, le64_to_cpu(raw_desc->m1)); in xge_start_xmit() 202 addr_lo = GET_BITS(NEXT_DESC_ADDRL, le64_to_cpu(raw_desc->m1)); in xge_start_xmit() 227 if (GET_BITS(E, le64_to_cpu(raw_desc->m0)) && in is_tx_hw_done() 228 !GET_BITS(PKT_SIZE, le64_to_cpu(raw_desc->m0))) in is_tx_hw_done() 307 if (GET_BITS(E, le64_to_cpu(raw_desc->m0))) in xge_rx_poll() 315 len = GET_BITS(PKT_SIZE, le64_to_cpu(raw_desc->m0)); in xge_rx_poll() [all …]
|
| /linux/fs/nilfs2/ |
| H A D | the_nilfs.c | 134 nilfs->ns_nongc_ctime = le64_to_cpu(raw_sr->sr_nongc_ctime); in nilfs_load_super_root() 177 nilfs->ns_last_pseg = le64_to_cpu(sbp->s_last_pseg); in nilfs_store_log_cursor() 178 nilfs->ns_last_cno = le64_to_cpu(sbp->s_last_cno); in nilfs_store_log_cursor() 179 nilfs->ns_last_seq = le64_to_cpu(sbp->s_last_seq); in nilfs_store_log_cursor() 280 nilfs->ns_sbwtime = le64_to_cpu(sbp[0]->s_wtime); in load_nilfs() 329 features = le64_to_cpu(nilfs->ns_sbp[0]->s_feature_compat_ro) & in load_nilfs() 480 nilfs->ns_first_data_block = le64_to_cpu(sbp->s_first_data_block); in nilfs_store_disk_layout() 491 nsegments = le64_to_cpu(sbp->s_nsegments); in nilfs_store_disk_layout() 560 u64 nsegments = le64_to_cpu(sbp->s_nsegments); in nilfs_sb2_bad_offset() 645 le64_to_cpu(sbp[1]->s_last_cno) > in nilfs_load_super_block() [all …]
|
| H A D | cpfile.c | 285 atomic64_set(&root->inodes_count, le64_to_cpu(cp->cp_inodes_count)); in nilfs_cpfile_read_checkpoint() 286 atomic64_set(&root->blocks_count, le64_to_cpu(cp->cp_blocks_count)); in nilfs_cpfile_read_checkpoint() 559 ci->ci_cno = le64_to_cpu(cp->cp_cno); in nilfs_cpfile_checkpoint_to_cpinfo() 560 ci->ci_create = le64_to_cpu(cp->cp_create); in nilfs_cpfile_checkpoint_to_cpinfo() 561 ci->ci_nblk_inc = le64_to_cpu(cp->cp_nblk_inc); in nilfs_cpfile_checkpoint_to_cpinfo() 562 ci->ci_inodes_count = le64_to_cpu(cp->cp_inodes_count); in nilfs_cpfile_checkpoint_to_cpinfo() 563 ci->ci_blocks_count = le64_to_cpu(cp->cp_blocks_count); in nilfs_cpfile_checkpoint_to_cpinfo() 564 ci->ci_next = le64_to_cpu(cp->cp_snapshot_list.ssl_next); in nilfs_cpfile_checkpoint_to_cpinfo() 640 curr = le64_to_cpu(header->ch_snapshot_list.ssl_next); in nilfs_cpfile_do_get_ssinfo() 669 next = le64_to_cpu(cp->cp_snapshot_list.ssl_next); in nilfs_cpfile_do_get_ssinfo() [all …]
|
| /linux/drivers/net/ethernet/intel/igc/ |
| H A D | igc_dump.c | 192 i, le64_to_cpu(u0->a), in igc_rings_dump() 193 le64_to_cpu(u0->b), in igc_rings_dump() 275 le64_to_cpu(u0->a), in igc_rings_dump() 276 le64_to_cpu(u0->b), in igc_rings_dump() 281 le64_to_cpu(u0->a), in igc_rings_dump() 282 le64_to_cpu(u0->b), in igc_rings_dump()
|
| /linux/drivers/md/ |
| H A D | dm-thin-metadata.c | 271 if (dm_block_location(b) != le64_to_cpu(disk_super->blocknr)) { in sb_check() 273 __func__, le64_to_cpu(disk_super->blocknr), in sb_check() 278 if (le64_to_cpu(disk_super->magic) != THIN_SUPERBLOCK_MAGIC) { in sb_check() 280 __func__, le64_to_cpu(disk_super->magic), in sb_check() 335 unpack_block_time(le64_to_cpu(*value_le), &b, &t); in with_runs() 376 unpack_block_time(le64_to_cpu(v1_le), &b1, &t); in data_block_equal() 377 unpack_block_time(le64_to_cpu(v2_le), &b2, &t); in data_block_equal() 389 dm_tm_inc(info->tm, le64_to_cpu(*root_le)); in subtree_inc() 399 if (dm_btree_del(info, le64_to_cpu(*root_le))) in subtree_dec() 749 pmd->root = le64_to_cpu(disk_super->data_mapping_root); in __open_metadata() [all …]
|
| /linux/drivers/net/ethernet/netronome/nfp/nfpcore/ |
| H A D | nfp_nsp_eth.c | 139 port = le64_to_cpu(src->port); in nfp_eth_port_translate() 140 state = le64_to_cpu(src->state); in nfp_eth_port_translate() 262 entry->link_modes_supp[i] = le64_to_cpu(ethm.supported_modes[i]); in nfp_eth_read_media() 263 entry->link_modes_ad[i] = le64_to_cpu(ethm.advertised_modes[i]); in nfp_eth_read_media() 450 reg = le64_to_cpu(entries[idx].state); in nfp_eth_set_mod_enable() 452 reg = le64_to_cpu(entries[idx].control); in nfp_eth_set_mod_enable() 497 reg = le64_to_cpu(entries[idx].state); in nfp_eth_set_configured() 499 reg = le64_to_cpu(entries[idx].control); in nfp_eth_set_configured() 529 reg = le64_to_cpu(entries[idx].raw[raw_idx]); in nfp_eth_set_bit_config() 564 reg = le64_to_cpu(entries[idx].control); in nfp_eth_set_idmode()
|
| /linux/arch/riscv/kernel/ |
| H A D | kexec_image.c | 55 flags = le64_to_cpu(h->flags); in image_load() 72 kbuf.memsz = le64_to_cpu(h->image_size); in image_load() 73 kbuf.buf_align = le64_to_cpu(h->text_offset); in image_load()
|
| /linux/fs/ubifs/ |
| H A D | master.c | 355 c->max_sqnum = le64_to_cpu(c->mst_node->ch.sqnum); in ubifs_read_master() 356 c->highest_inum = le64_to_cpu(c->mst_node->highest_inum); in ubifs_read_master() 357 c->cmt_no = le64_to_cpu(c->mst_node->cmt_no); in ubifs_read_master() 365 c->bi.old_idx_sz = le64_to_cpu(c->mst_node->index_size); in ubifs_read_master() 378 c->lst.total_free = le64_to_cpu(c->mst_node->total_free); in ubifs_read_master() 379 c->lst.total_dirty = le64_to_cpu(c->mst_node->total_dirty); in ubifs_read_master() 380 c->lst.total_used = le64_to_cpu(c->mst_node->total_used); in ubifs_read_master() 381 c->lst.total_dead = le64_to_cpu(c->mst_node->total_dead); in ubifs_read_master() 382 c->lst.total_dark = le64_to_cpu(c->mst_node->total_dark); in ubifs_read_master()
|
| /linux/drivers/bus/mhi/ep/ |
| H A D | ring.c | 21 return le64_to_cpu(rlen) / sizeof(struct mhi_ring_element); in mhi_ep_ring_num_elems() 189 ring->rbase = le64_to_cpu(val); in mhi_ep_ring_start() 204 ring->rd_offset = mhi_ep_ring_addr2offset(ring, le64_to_cpu(val)); in mhi_ep_ring_start() 205 ring->wr_offset = mhi_ep_ring_addr2offset(ring, le64_to_cpu(val)); in mhi_ep_ring_start() 214 ret = mhi_ep_cache_ring(ring, le64_to_cpu(val)); in mhi_ep_ring_start()
|
| /linux/drivers/md/persistent-data/ |
| H A D | dm-btree.c | 50 uint64_t mid_key = le64_to_cpu(n->keys[mid]); in bsearch() 367 *result_key = le64_to_cpu(ro_node(s)->keys[i]); in btree_lookup_raw() 411 root = le64_to_cpu(internal_value_le); in dm_btree_lookup() 462 *rkey = le64_to_cpu(n->keys[i]); in dm_btree_lookup_next_single() 491 root = le64_to_cpu(internal_value_le); in dm_btree_lookup_next() 684 le64_to_cpu(rn->keys[0]), &location); in split_one_into_two() 691 if (key < le64_to_cpu(rn->keys[0])) { in split_one_into_two() 783 le64_to_cpu(mn->keys[0]), &location); in split_two_into_three() 798 if (key < le64_to_cpu(mn->keys[0])) { in split_two_into_three() 802 } else if (key < le64_to_cpu(rn->keys[0])) { in split_two_into_three() [all …]
|
| /linux/drivers/iommu/arm/arm-smmu-v3/ |
| H A D | arm-smmu-v3-iommufd.c | 72 FIELD_GET(STRTAB_STE_0_CFG, le64_to_cpu(nested_domain->ste[0])); in arm_smmu_make_nested_domain_ste() 103 FIELD_GET(STRTAB_STE_0_CFG, le64_to_cpu(nested_domain->ste[0])); in arm_smmu_attach_prepare_vmaster() 180 if (FIELD_GET(STRTAB_STE_0_CFG, le64_to_cpu(nested_domain->ste[0])) == in arm_smmu_attach_dev_nested() 223 cfg = FIELD_GET(STRTAB_STE_0_CFG, le64_to_cpu(arg->ste[0])); in arm_smmu_validate_vste() 232 eats = FIELD_GET(STRTAB_STE_1_EATS, le64_to_cpu(arg->ste[1])); in arm_smmu_validate_vste() 319 cmd->cmd[0] = le64_to_cpu(cmd->ucmd.cmd[0]); in arm_vsmmu_convert_user_cmd() 320 cmd->cmd[1] = le64_to_cpu(cmd->ucmd.cmd[1]); in arm_vsmmu_convert_user_cmd()
|
| /linux/drivers/nvme/host/ |
| H A D | zns.c | 92 zi->zone_size = le64_to_cpu(id->lbafe[lbaf].zsze); in nvme_query_zone_info() 165 zone.capacity = nvme_lba_to_sect(head, le64_to_cpu(entry->zcap)); in nvme_zone_parse_entry() 166 zone.start = nvme_lba_to_sect(head, le64_to_cpu(entry->zslba)); in nvme_zone_parse_entry() 170 zone.wp = nvme_lba_to_sect(head, le64_to_cpu(entry->wp)); in nvme_zone_parse_entry() 210 nz = min((unsigned int)le64_to_cpu(report->nr_zones), nr_zones); in nvme_ns_report_zones()
|
| /linux/drivers/net/ethernet/qlogic/qlcnic/ |
| H A D | qlcnic_ctx.c | 1110 le64_to_cpu(stats->multicast_frames); in qlcnic_get_port_stats() 1112 le64_to_cpu(stats->broadcast_frames); in qlcnic_get_port_stats() 1113 esw_stats->unicast_frames = le64_to_cpu(stats->unicast_frames); in qlcnic_get_port_stats() 1114 esw_stats->dropped_frames = le64_to_cpu(stats->dropped_frames); in qlcnic_get_port_stats() 1115 esw_stats->local_frames = le64_to_cpu(stats->local_frames); in qlcnic_get_port_stats() 1116 esw_stats->errors = le64_to_cpu(stats->errors); in qlcnic_get_port_stats() 1117 esw_stats->numbytes = le64_to_cpu(stats->numbytes); in qlcnic_get_port_stats() 1157 mac_stats->mac_tx_frames = le64_to_cpu(stats->mac_tx_frames); in qlcnic_get_mac_stats() 1158 mac_stats->mac_tx_bytes = le64_to_cpu(stats->mac_tx_bytes); in qlcnic_get_mac_stats() 1160 le64_to_cpu(stats->mac_tx_mcast_pkts); in qlcnic_get_mac_stats() [all …]
|
| /linux/drivers/scsi/megaraid/ |
| H A D | megaraid_sas_fp.c | 388 if (le64_to_cpu(quad->logStart) <= row && row <= in MR_GetSpanBlock() 389 le64_to_cpu(quad->logEnd) && (mega_mod64(row - le64_to_cpu(quad->logStart), in MR_GetSpanBlock() 393 blk = mega_div64_32((row-le64_to_cpu(quad->logStart)), le32_to_cpu(quad->diff)); in MR_GetSpanBlock() 395 blk = (blk + le64_to_cpu(quad->offsetInSpan)) << raid->stripeShift; in MR_GetSpanBlock() 450 if (le64_to_cpu(quad->logStart) <= row && in mr_spanset_get_span_block() 451 row <= le64_to_cpu(quad->logEnd) && in mr_spanset_get_span_block() 452 (mega_mod64(row - le64_to_cpu(quad->logStart), in mr_spanset_get_span_block() 457 ((row - le64_to_cpu(quad->logStart)), in mr_spanset_get_span_block() 459 blk = (blk + le64_to_cpu(quad->offsetInSpan)) in mr_spanset_get_span_block() 567 if (le64_to_cpu(quad->logStart) <= row && in get_strip_from_row() [all …]
|
| /linux/include/soc/fsl/ |
| H A D | dpaa2-fd.h | 106 return (dma_addr_t)le64_to_cpu(fd->simple.addr); in dpaa2_fd_get_addr() 169 return (dma_addr_t)le64_to_cpu(fd->simple.flc); in dpaa2_fd_get_flc() 312 return (dma_addr_t)le64_to_cpu(sg->addr); in dpaa2_sg_get_addr() 482 return (dma_addr_t)le64_to_cpu(fle->addr); in dpaa2_fl_get_addr() 546 return (dma_addr_t)le64_to_cpu(fle->flc); in dpaa2_fl_get_flc()
|