/* * kmp_taskdeps.h */ //===----------------------------------------------------------------------===// // // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. // See https://llvm.org/LICENSE.txt for license information. // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception // //===----------------------------------------------------------------------===// #ifndef KMP_TASKDEPS_H #define KMP_TASKDEPS_H #include "kmp.h" #define KMP_ACQUIRE_DEPNODE(gtid, n) __kmp_acquire_lock(&(n)->dn.lock, (gtid)) #define KMP_RELEASE_DEPNODE(gtid, n) __kmp_release_lock(&(n)->dn.lock, (gtid)) static inline void __kmp_node_deref(kmp_info_t *thread, kmp_depnode_t *node) { if (!node) return; kmp_int32 n = KMP_ATOMIC_DEC(&node->dn.nrefs) - 1; KMP_DEBUG_ASSERT(n >= 0); if (n == 0) { #if USE_ITT_BUILD && USE_ITT_NOTIFY __itt_sync_destroy(node); #endif KMP_ASSERT(node->dn.nrefs == 0); #if USE_FAST_MEMORY __kmp_fast_free(thread, node); #else __kmp_thread_free(thread, node); #endif } } static inline void __kmp_depnode_list_free(kmp_info_t *thread, kmp_depnode_list *list) { kmp_depnode_list *next; for (; list; list = next) { next = list->next; __kmp_node_deref(thread, list->node); #if USE_FAST_MEMORY __kmp_fast_free(thread, list); #else __kmp_thread_free(thread, list); #endif } } static inline void __kmp_dephash_free_entries(kmp_info_t *thread, kmp_dephash_t *h) { for (size_t i = 0; i < h->size; i++) { if (h->buckets[i]) { kmp_dephash_entry_t *next; for (kmp_dephash_entry_t *entry = h->buckets[i]; entry; entry = next) { next = entry->next_in_bucket; __kmp_depnode_list_free(thread, entry->last_set); __kmp_depnode_list_free(thread, entry->prev_set); __kmp_node_deref(thread, entry->last_out); if (entry->mtx_lock) { __kmp_destroy_lock(entry->mtx_lock); __kmp_free(entry->mtx_lock); } #if USE_FAST_MEMORY __kmp_fast_free(thread, entry); #else __kmp_thread_free(thread, entry); #endif } h->buckets[i] = 0; } } __kmp_node_deref(thread, h->last_all); h->last_all = NULL; } static inline void __kmp_dephash_free(kmp_info_t *thread, kmp_dephash_t *h) { __kmp_dephash_free_entries(thread, h); #if USE_FAST_MEMORY __kmp_fast_free(thread, h); #else __kmp_thread_free(thread, h); #endif } extern void __kmpc_give_task(kmp_task_t *ptask, kmp_int32 start); static inline void __kmp_release_deps(kmp_int32 gtid, kmp_taskdata_t *task) { #if OMPX_TASKGRAPH if (task->is_taskgraph && !(__kmp_tdg_is_recording(task->tdg->tdg_status))) { kmp_node_info_t *TaskInfo = &(task->tdg->record_map[task->td_task_id]); for (int i = 0; i < TaskInfo->nsuccessors; i++) { kmp_int32 successorNumber = TaskInfo->successors[i]; kmp_node_info_t *successor = &(task->tdg->record_map[successorNumber]); kmp_int32 npredecessors = KMP_ATOMIC_DEC(&successor->npredecessors_counter) - 1; if (successor->task != nullptr && npredecessors == 0) { __kmp_omp_task(gtid, successor->task, false); } } return; } #endif kmp_info_t *thread = __kmp_threads[gtid]; kmp_depnode_t *node = task->td_depnode; // Check mutexinoutset dependencies, release locks if (UNLIKELY(node && (node->dn.mtx_num_locks < 0))) { // negative num_locks means all locks were acquired node->dn.mtx_num_locks = -node->dn.mtx_num_locks; for (int i = node->dn.mtx_num_locks - 1; i >= 0; --i) { KMP_DEBUG_ASSERT(node->dn.mtx_locks[i] != NULL); __kmp_release_lock(node->dn.mtx_locks[i], gtid); } } if (task->td_dephash) { KA_TRACE( 40, ("__kmp_release_deps: T#%d freeing dependencies hash of task %p.\n", gtid, task)); __kmp_dephash_free(thread, task->td_dephash); task->td_dephash = NULL; } if (!node) return; KA_TRACE(20, ("__kmp_release_deps: T#%d notifying successors of task %p.\n", gtid, task)); KMP_ACQUIRE_DEPNODE(gtid, node); #if OMPX_TASKGRAPH if (!task->is_taskgraph || (task->is_taskgraph && !__kmp_tdg_is_recording(task->tdg->tdg_status))) #endif node->dn.task = NULL; // mark this task as finished, so no new dependencies are generated KMP_RELEASE_DEPNODE(gtid, node); kmp_depnode_list_t *next; kmp_taskdata_t *next_taskdata; for (kmp_depnode_list_t *p = node->dn.successors; p; p = next) { kmp_depnode_t *successor = p->node; #if USE_ITT_BUILD && USE_ITT_NOTIFY __itt_sync_releasing(successor); #endif kmp_int32 npredecessors = KMP_ATOMIC_DEC(&successor->dn.npredecessors) - 1; // successor task can be NULL for wait_depends or because deps are still // being processed if (npredecessors == 0) { #if USE_ITT_BUILD && USE_ITT_NOTIFY __itt_sync_acquired(successor); #endif KMP_MB(); if (successor->dn.task) { KA_TRACE(20, ("__kmp_release_deps: T#%d successor %p of %p scheduled " "for execution.\n", gtid, successor->dn.task, task)); // If a regular task depending on a hidden helper task, when the // hidden helper task is done, the regular task should be executed by // its encountering team. if (KMP_HIDDEN_HELPER_THREAD(gtid)) { // Hidden helper thread can only execute hidden helper tasks KMP_ASSERT(task->td_flags.hidden_helper); next_taskdata = KMP_TASK_TO_TASKDATA(successor->dn.task); // If the dependent task is a regular task, we need to push to its // encountering thread's queue; otherwise, it can be pushed to its own // queue. if (!next_taskdata->td_flags.hidden_helper) { kmp_int32 encountering_gtid = next_taskdata->td_alloc_thread->th.th_info.ds.ds_gtid; kmp_int32 encountering_tid = __kmp_tid_from_gtid(encountering_gtid); __kmpc_give_task(successor->dn.task, encountering_tid); } else { __kmp_omp_task(gtid, successor->dn.task, false); } } else { __kmp_omp_task(gtid, successor->dn.task, false); } } } next = p->next; __kmp_node_deref(thread, p->node); #if USE_FAST_MEMORY __kmp_fast_free(thread, p); #else __kmp_thread_free(thread, p); #endif } __kmp_node_deref(thread, node); KA_TRACE( 20, ("__kmp_release_deps: T#%d all successors of %p notified of completion\n", gtid, task)); } #endif // KMP_TASKDEPS_H