Lines Matching refs:TIF_PATCH_PENDING
123 WARN_ON_ONCE(test_tsk_thread_flag(task, TIF_PATCH_PENDING));
130 WARN_ON_ONCE(test_tsk_thread_flag(task, TIF_PATCH_PENDING));
187 * 1) Enforce the order of the TIF_PATCH_PENDING read and the
191 * 2) Enforce the order of the TIF_PATCH_PENDING read and a future read
195 if (test_and_clear_tsk_thread_flag(task, TIF_PATCH_PENDING))
295 clear_tsk_thread_flag(task, TIF_PATCH_PENDING);
373 * Enforce the order of the TIF_PATCH_PENDING read above and the
384 * Sends a fake signal to all non-kthread tasks with TIF_PATCH_PENDING set.
385 * Kthreads with TIF_PATCH_PENDING set are woken up.
400 * There is a small race here. We could see TIF_PATCH_PENDING
468 clear_tsk_thread_flag(task, TIF_PATCH_PENDING);
528 set_tsk_thread_flag(task, TIF_PATCH_PENDING);
539 set_tsk_thread_flag(task, TIF_PATCH_PENDING);
566 * has no effect until the TIF_PATCH_PENDING flags get set later.
599 * TIF_PATCH_PENDING writes to ensure klp_update_patch_state() and
638 * Clear all TIF_PATCH_PENDING flags to prevent races caused by
644 clear_tsk_thread_flag(task, TIF_PATCH_PENDING);
648 clear_tsk_thread_flag(idle_task(cpu), TIF_PATCH_PENDING);
652 * __klp_sched_try_switch() see the cleared TIF_PATCH_PENDING before
666 * TIF_PATCH_PENDING writes in klp_start_transition() to ensure
689 if (test_tsk_thread_flag(current, TIF_PATCH_PENDING))
690 set_tsk_thread_flag(child, TIF_PATCH_PENDING);
692 clear_tsk_thread_flag(child, TIF_PATCH_PENDING);
698 * Drop TIF_PATCH_PENDING of all tasks on admin's request. This forces an