/linux/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_fence.c | 123 seq = atomic_read(&drv->last_seq); in amdgpu_fence_read() 275 uint32_t seq, last_seq; in amdgpu_fence_process() local 278 last_seq = atomic_read(&ring->fence_drv.last_seq); in amdgpu_fence_process() 281 } while (atomic_cmpxchg(&drv->last_seq, last_seq, seq) != last_seq); in amdgpu_fence_process() 287 if (unlikely(seq == last_seq)) in amdgpu_fence_process() 290 last_seq &= drv->num_fences_mask; in amdgpu_fence_process() 296 ++last_seq; in amdgpu_fence_process() 297 last_seq &= drv->num_fences_mask; in amdgpu_fence_process() 298 ptr = &drv->fences[last_seq]; in amdgpu_fence_process() 311 } while (last_seq != seq); in amdgpu_fence_process() [all …]
|
H A D | amdgpu_ring_mux.c | 80 uint32_t seq, last_seq; in amdgpu_mux_resubmit_chunks() local 99 last_seq = atomic_read(&e->ring->fence_drv.last_seq); in amdgpu_mux_resubmit_chunks() 101 if (last_seq < seq) { in amdgpu_mux_resubmit_chunks() 104 if (chunk->sync_seq > last_seq && chunk->sync_seq <= seq) { in amdgpu_mux_resubmit_chunks() 463 uint32_t last_seq = 0; in scan_and_remove_signaled_chunk() local 473 last_seq = atomic_read(&ring->fence_drv.last_seq); in scan_and_remove_signaled_chunk() 476 if (chunk->sync_seq <= last_seq) { in scan_and_remove_signaled_chunk()
|
H A D | amdgpu_debugfs.c | 1818 uint32_t sync_seq, last_seq; in amdgpu_ib_preempt_fences_swap() local 1820 last_seq = atomic_read(&ring->fence_drv.last_seq); in amdgpu_ib_preempt_fences_swap() 1823 last_seq &= drv->num_fences_mask; in amdgpu_ib_preempt_fences_swap() 1829 ++last_seq; in amdgpu_ib_preempt_fences_swap() 1830 last_seq &= drv->num_fences_mask; in amdgpu_ib_preempt_fences_swap() 1831 ptr = &drv->fences[last_seq]; in amdgpu_ib_preempt_fences_swap() 1839 fences[last_seq] = fence; in amdgpu_ib_preempt_fences_swap() 1841 } while (last_seq != sync_seq); in amdgpu_ib_preempt_fences_swap() 1886 if (preempt_seq <= atomic_read(&drv->last_seq)) { in amdgpu_ib_preempt_mark_partial_job() 1954 if (atomic_read(&ring->fence_drv.last_seq) != in amdgpu_debugfs_ib_preempt()
|
H A D | amdgpu_job.c | 124 job->base.sched->name, atomic_read(&ring->fence_drv.last_seq), in amdgpu_job_timedout()
|
H A D | amdgpu_ring.h | 119 atomic_t last_seq; member
|
/linux/drivers/gpu/drm/radeon/ |
H A D | radeon_fence.c | 97 seq = lower_32_bits(atomic64_read(&drv->last_seq)); in radeon_fence_read() 177 seq = atomic64_read(&fence->rdev->fence_drv[fence->ring].last_seq); in radeon_fence_check_signaled() 199 uint64_t seq, last_seq, last_emitted; in radeon_fence_activity() local 224 last_seq = atomic64_read(&rdev->fence_drv[ring].last_seq); in radeon_fence_activity() 228 seq |= last_seq & 0xffffffff00000000LL; in radeon_fence_activity() 229 if (seq < last_seq) { in radeon_fence_activity() 234 if (seq <= last_seq || seq > last_emitted) in radeon_fence_activity() 242 last_seq = seq; in radeon_fence_activity() 251 } while (atomic64_xchg(&rdev->fence_drv[ring].last_seq, seq) > seq); in radeon_fence_activity() 300 (uint64_t)atomic64_read(&fence_drv->last_seq), in radeon_fence_check_lockup() [all …]
|
/linux/fs/bcachefs/ |
H A D | journal_io.c | 125 u64 last_seq; member 144 u64 last_seq = !JSET_NO_FLUSH(j) ? le64_to_cpu(j->last_seq) : 0; in journal_entry_add() local 154 le64_to_cpu(j->seq) < jlist->last_seq) in journal_entry_add() 166 if (last_seq > jlist->last_seq && !c->opts.read_entire_journal) { in journal_entry_add() 168 journal_entry_radix_idx(c, jlist->last_seq)) { in journal_entry_add() 174 if (le64_to_cpu(i->j.seq) >= last_seq) in journal_entry_add() 181 jlist->last_seq = max(jlist->last_seq, last_seq); in journal_entry_add() 908 le64_to_cpu(jset->last_seq) > le64_to_cpu(jset->seq), in jset_validate() 912 le64_to_cpu(jset->last_seq), in jset_validate() 914 jset->last_seq = jset->seq; in jset_validate() [all …]
|
H A D | journal.c | 293 buf->last_seq = journal_last_seq(j); in __journal_entry_close() 294 buf->data->last_seq = cpu_to_le64(buf->last_seq); in __journal_entry_close() 295 BUG_ON(buf->last_seq > le64_to_cpu(buf->data->seq)); in __journal_entry_close() 1218 u64 last_seq = cur_seq, nr, seq; in bch2_fs_journal_start() local 1226 last_seq = le64_to_cpu(i->j.last_seq); in bch2_fs_journal_start() 1230 nr = cur_seq - last_seq; in bch2_fs_journal_start() 1241 j->replay_journal_seq = last_seq; in bch2_fs_journal_start() 1243 j->last_seq_ondisk = last_seq; in bch2_fs_journal_start() 1246 j->pin.front = last_seq; in bch2_fs_journal_start() 1262 if (seq < last_seq) in bch2_fs_journal_start()
|
H A D | recovery.c | 635 u64 last_seq = 0, blacklist_seq, journal_seq; in bch2_fs_recovery() local 736 ret = bch2_journal_read(c, &last_seq, &blacklist_seq, &journal_seq); in bch2_fs_recovery() 781 (*i)->j.last_seq = (*i)->j.seq; in bch2_fs_recovery() 807 c->journal_replay_seq_start = last_seq; in bch2_fs_recovery() 841 journal_seq, last_seq, blacklist_seq - 1) ?: in bch2_fs_recovery()
|
H A D | journal.h | 241 if (j->seq != j->last_seq) in journal_entry_empty()
|
H A D | bcachefs_format.h | 1255 __le64 last_seq; member
|
/linux/drivers/md/bcache/ |
H A D | journal.c | 118 if (i->j.seq >= j->last_seq) in journal_read_bucket() 133 if (j->seq < i->j.last_seq) in journal_read_bucket() 356 uint64_t start = i->j.last_seq, end = i->j.seq, n = start; in bch_journal_replay() 569 #define last_seq(j) ((j)->seq - fifo_used(&(j)->pin) + 1) macro 653 uint64_t last_seq; in journal_reclaim() local 662 last_seq = last_seq(&c->journal); in journal_reclaim() 667 ja->seq[ja->last_idx] < last_seq) in journal_reclaim() 785 w->data->last_seq = last_seq(&c->journal);
|
H A D | bcache_ondisk.h | 353 __u64 last_seq; member
|
/linux/drivers/net/wireless/marvell/mwifiex/ |
H A D | 11n_rxreorder.c | 334 u16 last_seq = 0; in mwifiex_11n_create_rx_reorder_tbl() local 363 last_seq = node->rx_seq[tid]; in mwifiex_11n_create_rx_reorder_tbl() 368 last_seq = node->rx_seq[tid]; in mwifiex_11n_create_rx_reorder_tbl() 370 last_seq = priv->rx_seq[tid]; in mwifiex_11n_create_rx_reorder_tbl() 376 last_seq, new_node->start_win); in mwifiex_11n_create_rx_reorder_tbl() 378 if (last_seq != MWIFIEX_DEF_11N_RX_SEQ_NUM && in mwifiex_11n_create_rx_reorder_tbl() 379 last_seq >= new_node->start_win) { in mwifiex_11n_create_rx_reorder_tbl() 380 new_node->start_win = last_seq + 1; in mwifiex_11n_create_rx_reorder_tbl()
|
/linux/include/linux/netfilter/ |
H A D | nf_conntrack_tcp.h | 24 u_int32_t last_seq; /* Last sequence number seen in dir */ member
|
/linux/net/rxrpc/ |
H A D | conn_event.c | 156 pkt.ack.firstPacket = htonl(chan->last_seq + 1); in rxrpc_conn_retransmit_call() 157 pkt.ack.previousPacket = htonl(chan->last_seq); in rxrpc_conn_retransmit_call()
|
H A D | ar-internal.h | 493 u32 last_seq; member
|
/linux/include/net/ |
H A D | mctp.h | 162 u8 last_seq; member
|
/linux/fs/nilfs2/ |
H A D | sysfs.c | 464 u64 last_seq; in nilfs_segctor_last_seg_sequence_show() local 467 last_seq = nilfs->ns_last_seq; in nilfs_segctor_last_seg_sequence_show() 470 return sysfs_emit(buf, "%llu\n", last_seq); in nilfs_segctor_last_seg_sequence_show()
|
/linux/fs/btrfs/ |
H A D | tree-checker.c | 1317 u64 last_seq = U64_MAX; in check_extent_item() local 1570 last_seq = U64_MAX; in check_extent_item() 1571 if (seq > last_seq) { in check_extent_item() 1575 last_type, last_seq); in check_extent_item() 1579 last_seq = seq; in check_extent_item()
|
/linux/net/mctp/ |
H A D | route.c | 337 key->last_seq = this_seq; in mctp_frag_queue() 341 exp_seq = (key->last_seq + 1) & MCTP_HDR_SEQ_MASK; in mctp_frag_queue() 354 key->last_seq = this_seq; in mctp_frag_queue()
|
/linux/net/netfilter/ |
H A D | nf_conntrack_proto_tcp.c | 704 state->last_seq == seq && in tcp_in_window() 711 state->last_seq = seq; in tcp_in_window() 1072 ct->proto.tcp.last_seq = ntohl(th->seq); in nf_conntrack_tcp_packet()
|
/linux/drivers/net/wireless/intel/iwlwifi/mvm/ |
H A D | sta.h | 322 __le16 last_seq[IWL_MAX_TID_COUNT + 1]; member
|
H A D | rxmq.c | 536 dup_data->last_seq[tid] == hdr->seq_ctrl && in iwl_mvm_is_dup() 541 if (dup_data->last_seq[tid] == hdr->seq_ctrl && in iwl_mvm_is_dup() 546 dup_data->last_seq[tid] = hdr->seq_ctrl; in iwl_mvm_is_dup()
|
/linux/drivers/scsi/lpfc/ |
H A D | lpfc.h | 1003 u8 last_seq, u8 cr_cx_cmd);
|