Home
last modified time | relevance | path

Searched refs:last_seq (Results 1 – 25 of 27) sorted by relevance

12

/linux/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_fence.c123 seq = atomic_read(&drv->last_seq); in amdgpu_fence_read()
275 uint32_t seq, last_seq; in amdgpu_fence_process() local
278 last_seq = atomic_read(&ring->fence_drv.last_seq); in amdgpu_fence_process()
281 } while (atomic_cmpxchg(&drv->last_seq, last_seq, seq) != last_seq); in amdgpu_fence_process()
287 if (unlikely(seq == last_seq)) in amdgpu_fence_process()
290 last_seq &= drv->num_fences_mask; in amdgpu_fence_process()
296 ++last_seq; in amdgpu_fence_process()
297 last_seq &= drv->num_fences_mask; in amdgpu_fence_process()
298 ptr = &drv->fences[last_seq]; in amdgpu_fence_process()
311 } while (last_seq != seq); in amdgpu_fence_process()
[all …]
H A Damdgpu_ring_mux.c80 uint32_t seq, last_seq; in amdgpu_mux_resubmit_chunks() local
99 last_seq = atomic_read(&e->ring->fence_drv.last_seq); in amdgpu_mux_resubmit_chunks()
101 if (last_seq < seq) { in amdgpu_mux_resubmit_chunks()
104 if (chunk->sync_seq > last_seq && chunk->sync_seq <= seq) { in amdgpu_mux_resubmit_chunks()
463 uint32_t last_seq = 0; in scan_and_remove_signaled_chunk() local
473 last_seq = atomic_read(&ring->fence_drv.last_seq); in scan_and_remove_signaled_chunk()
476 if (chunk->sync_seq <= last_seq) { in scan_and_remove_signaled_chunk()
H A Damdgpu_debugfs.c1818 uint32_t sync_seq, last_seq; in amdgpu_ib_preempt_fences_swap() local
1820 last_seq = atomic_read(&ring->fence_drv.last_seq); in amdgpu_ib_preempt_fences_swap()
1823 last_seq &= drv->num_fences_mask; in amdgpu_ib_preempt_fences_swap()
1829 ++last_seq; in amdgpu_ib_preempt_fences_swap()
1830 last_seq &= drv->num_fences_mask; in amdgpu_ib_preempt_fences_swap()
1831 ptr = &drv->fences[last_seq]; in amdgpu_ib_preempt_fences_swap()
1839 fences[last_seq] = fence; in amdgpu_ib_preempt_fences_swap()
1841 } while (last_seq != sync_seq); in amdgpu_ib_preempt_fences_swap()
1886 if (preempt_seq <= atomic_read(&drv->last_seq)) { in amdgpu_ib_preempt_mark_partial_job()
1954 if (atomic_read(&ring->fence_drv.last_seq) != in amdgpu_debugfs_ib_preempt()
H A Damdgpu_job.c124 job->base.sched->name, atomic_read(&ring->fence_drv.last_seq), in amdgpu_job_timedout()
H A Damdgpu_ring.h119 atomic_t last_seq; member
/linux/drivers/gpu/drm/radeon/
H A Dradeon_fence.c97 seq = lower_32_bits(atomic64_read(&drv->last_seq)); in radeon_fence_read()
177 seq = atomic64_read(&fence->rdev->fence_drv[fence->ring].last_seq); in radeon_fence_check_signaled()
199 uint64_t seq, last_seq, last_emitted; in radeon_fence_activity() local
224 last_seq = atomic64_read(&rdev->fence_drv[ring].last_seq); in radeon_fence_activity()
228 seq |= last_seq & 0xffffffff00000000LL; in radeon_fence_activity()
229 if (seq < last_seq) { in radeon_fence_activity()
234 if (seq <= last_seq || seq > last_emitted) in radeon_fence_activity()
242 last_seq = seq; in radeon_fence_activity()
251 } while (atomic64_xchg(&rdev->fence_drv[ring].last_seq, seq) > seq); in radeon_fence_activity()
300 (uint64_t)atomic64_read(&fence_drv->last_seq), in radeon_fence_check_lockup()
[all …]
/linux/fs/bcachefs/
H A Djournal_io.c125 u64 last_seq; member
144 u64 last_seq = !JSET_NO_FLUSH(j) ? le64_to_cpu(j->last_seq) : 0; in journal_entry_add() local
154 le64_to_cpu(j->seq) < jlist->last_seq) in journal_entry_add()
166 if (last_seq > jlist->last_seq && !c->opts.read_entire_journal) { in journal_entry_add()
168 journal_entry_radix_idx(c, jlist->last_seq)) { in journal_entry_add()
174 if (le64_to_cpu(i->j.seq) >= last_seq) in journal_entry_add()
181 jlist->last_seq = max(jlist->last_seq, last_seq); in journal_entry_add()
908 le64_to_cpu(jset->last_seq) > le64_to_cpu(jset->seq), in jset_validate()
912 le64_to_cpu(jset->last_seq), in jset_validate()
914 jset->last_seq = jset->seq; in jset_validate()
[all …]
H A Djournal.c293 buf->last_seq = journal_last_seq(j); in __journal_entry_close()
294 buf->data->last_seq = cpu_to_le64(buf->last_seq); in __journal_entry_close()
295 BUG_ON(buf->last_seq > le64_to_cpu(buf->data->seq)); in __journal_entry_close()
1218 u64 last_seq = cur_seq, nr, seq; in bch2_fs_journal_start() local
1226 last_seq = le64_to_cpu(i->j.last_seq); in bch2_fs_journal_start()
1230 nr = cur_seq - last_seq; in bch2_fs_journal_start()
1241 j->replay_journal_seq = last_seq; in bch2_fs_journal_start()
1243 j->last_seq_ondisk = last_seq; in bch2_fs_journal_start()
1246 j->pin.front = last_seq; in bch2_fs_journal_start()
1262 if (seq < last_seq) in bch2_fs_journal_start()
H A Drecovery.c635 u64 last_seq = 0, blacklist_seq, journal_seq; in bch2_fs_recovery() local
736 ret = bch2_journal_read(c, &last_seq, &blacklist_seq, &journal_seq); in bch2_fs_recovery()
781 (*i)->j.last_seq = (*i)->j.seq; in bch2_fs_recovery()
807 c->journal_replay_seq_start = last_seq; in bch2_fs_recovery()
841 journal_seq, last_seq, blacklist_seq - 1) ?: in bch2_fs_recovery()
H A Djournal.h241 if (j->seq != j->last_seq) in journal_entry_empty()
H A Dbcachefs_format.h1255 __le64 last_seq; member
/linux/drivers/md/bcache/
H A Djournal.c118 if (i->j.seq >= j->last_seq) in journal_read_bucket()
133 if (j->seq < i->j.last_seq) in journal_read_bucket()
356 uint64_t start = i->j.last_seq, end = i->j.seq, n = start; in bch_journal_replay()
569 #define last_seq(j) ((j)->seq - fifo_used(&(j)->pin) + 1) macro
653 uint64_t last_seq; in journal_reclaim() local
662 last_seq = last_seq(&c->journal); in journal_reclaim()
667 ja->seq[ja->last_idx] < last_seq) in journal_reclaim()
785 w->data->last_seq = last_seq(&c->journal);
H A Dbcache_ondisk.h353 __u64 last_seq; member
/linux/drivers/net/wireless/marvell/mwifiex/
H A D11n_rxreorder.c334 u16 last_seq = 0; in mwifiex_11n_create_rx_reorder_tbl() local
363 last_seq = node->rx_seq[tid]; in mwifiex_11n_create_rx_reorder_tbl()
368 last_seq = node->rx_seq[tid]; in mwifiex_11n_create_rx_reorder_tbl()
370 last_seq = priv->rx_seq[tid]; in mwifiex_11n_create_rx_reorder_tbl()
376 last_seq, new_node->start_win); in mwifiex_11n_create_rx_reorder_tbl()
378 if (last_seq != MWIFIEX_DEF_11N_RX_SEQ_NUM && in mwifiex_11n_create_rx_reorder_tbl()
379 last_seq >= new_node->start_win) { in mwifiex_11n_create_rx_reorder_tbl()
380 new_node->start_win = last_seq + 1; in mwifiex_11n_create_rx_reorder_tbl()
/linux/include/linux/netfilter/
H A Dnf_conntrack_tcp.h24 u_int32_t last_seq; /* Last sequence number seen in dir */ member
/linux/net/rxrpc/
H A Dconn_event.c156 pkt.ack.firstPacket = htonl(chan->last_seq + 1); in rxrpc_conn_retransmit_call()
157 pkt.ack.previousPacket = htonl(chan->last_seq); in rxrpc_conn_retransmit_call()
H A Dar-internal.h493 u32 last_seq; member
/linux/include/net/
H A Dmctp.h162 u8 last_seq; member
/linux/fs/nilfs2/
H A Dsysfs.c464 u64 last_seq; in nilfs_segctor_last_seg_sequence_show() local
467 last_seq = nilfs->ns_last_seq; in nilfs_segctor_last_seg_sequence_show()
470 return sysfs_emit(buf, "%llu\n", last_seq); in nilfs_segctor_last_seg_sequence_show()
/linux/fs/btrfs/
H A Dtree-checker.c1317 u64 last_seq = U64_MAX; in check_extent_item() local
1570 last_seq = U64_MAX; in check_extent_item()
1571 if (seq > last_seq) { in check_extent_item()
1575 last_type, last_seq); in check_extent_item()
1579 last_seq = seq; in check_extent_item()
/linux/net/mctp/
H A Droute.c337 key->last_seq = this_seq; in mctp_frag_queue()
341 exp_seq = (key->last_seq + 1) & MCTP_HDR_SEQ_MASK; in mctp_frag_queue()
354 key->last_seq = this_seq; in mctp_frag_queue()
/linux/net/netfilter/
H A Dnf_conntrack_proto_tcp.c704 state->last_seq == seq && in tcp_in_window()
711 state->last_seq = seq; in tcp_in_window()
1072 ct->proto.tcp.last_seq = ntohl(th->seq); in nf_conntrack_tcp_packet()
/linux/drivers/net/wireless/intel/iwlwifi/mvm/
H A Dsta.h322 __le16 last_seq[IWL_MAX_TID_COUNT + 1]; member
H A Drxmq.c536 dup_data->last_seq[tid] == hdr->seq_ctrl && in iwl_mvm_is_dup()
541 if (dup_data->last_seq[tid] == hdr->seq_ctrl && in iwl_mvm_is_dup()
546 dup_data->last_seq[tid] = hdr->seq_ctrl; in iwl_mvm_is_dup()
/linux/drivers/scsi/lpfc/
H A Dlpfc.h1003 u8 last_seq, u8 cr_cx_cmd);

12