1 // SPDX-License-Identifier: GPL-2.0-only 2 3 /* 4 * A simple wrapper around refcount. An allocated sched_core_cookie's 5 * address is used to compute the cookie of the task. 6 */ 7 struct sched_core_cookie { 8 refcount_t refcnt; 9 }; 10 11 static unsigned long sched_core_alloc_cookie(void) 12 { 13 struct sched_core_cookie *ck = kmalloc(sizeof(*ck), GFP_KERNEL); 14 if (!ck) 15 return 0; 16 17 refcount_set(&ck->refcnt, 1); 18 sched_core_get(); 19 20 return (unsigned long)ck; 21 } 22 23 static void sched_core_put_cookie(unsigned long cookie) 24 { 25 struct sched_core_cookie *ptr = (void *)cookie; 26 27 if (ptr && refcount_dec_and_test(&ptr->refcnt)) { 28 kfree(ptr); 29 sched_core_put(); 30 } 31 } 32 33 static unsigned long sched_core_get_cookie(unsigned long cookie) 34 { 35 struct sched_core_cookie *ptr = (void *)cookie; 36 37 if (ptr) 38 refcount_inc(&ptr->refcnt); 39 40 return cookie; 41 } 42 43 /* 44 * sched_core_update_cookie - replace the cookie on a task 45 * @p: the task to update 46 * @cookie: the new cookie 47 * 48 * Effectively exchange the task cookie; caller is responsible for lifetimes on 49 * both ends. 50 * 51 * Returns: the old cookie 52 */ 53 static unsigned long sched_core_update_cookie(struct task_struct *p, 54 unsigned long cookie) 55 { 56 unsigned long old_cookie; 57 struct rq_flags rf; 58 struct rq *rq; 59 60 rq = task_rq_lock(p, &rf); 61 62 /* 63 * Since creating a cookie implies sched_core_get(), and we cannot set 64 * a cookie until after we've created it, similarly, we cannot destroy 65 * a cookie until after we've removed it, we must have core scheduling 66 * enabled here. 67 */ 68 SCHED_WARN_ON((p->core_cookie || cookie) && !sched_core_enabled(rq)); 69 70 if (sched_core_enqueued(p)) 71 sched_core_dequeue(rq, p, DEQUEUE_SAVE); 72 73 old_cookie = p->core_cookie; 74 p->core_cookie = cookie; 75 76 /* 77 * Consider the cases: !prev_cookie and !cookie. 78 */ 79 if (cookie && task_on_rq_queued(p)) 80 sched_core_enqueue(rq, p); 81 82 /* 83 * If task is currently running, it may not be compatible anymore after 84 * the cookie change, so enter the scheduler on its CPU to schedule it 85 * away. 86 * 87 * Note that it is possible that as a result of this cookie change, the 88 * core has now entered/left forced idle state. Defer accounting to the 89 * next scheduling edge, rather than always forcing a reschedule here. 90 */ 91 if (task_on_cpu(rq, p)) 92 resched_curr(rq); 93 94 task_rq_unlock(rq, p, &rf); 95 96 return old_cookie; 97 } 98 99 static unsigned long sched_core_clone_cookie(struct task_struct *p) 100 { 101 unsigned long cookie, flags; 102 103 raw_spin_lock_irqsave(&p->pi_lock, flags); 104 cookie = sched_core_get_cookie(p->core_cookie); 105 raw_spin_unlock_irqrestore(&p->pi_lock, flags); 106 107 return cookie; 108 } 109 110 void sched_core_fork(struct task_struct *p) 111 { 112 RB_CLEAR_NODE(&p->core_node); 113 p->core_cookie = sched_core_clone_cookie(current); 114 } 115 116 void sched_core_free(struct task_struct *p) 117 { 118 sched_core_put_cookie(p->core_cookie); 119 } 120 121 static void __sched_core_set(struct task_struct *p, unsigned long cookie) 122 { 123 cookie = sched_core_get_cookie(cookie); 124 cookie = sched_core_update_cookie(p, cookie); 125 sched_core_put_cookie(cookie); 126 } 127 128 /* Called from prctl interface: PR_SCHED_CORE */ 129 int sched_core_share_pid(unsigned int cmd, pid_t pid, enum pid_type type, 130 unsigned long uaddr) 131 { 132 unsigned long cookie = 0, id = 0; 133 struct task_struct *task, *p; 134 struct pid *grp; 135 int err = 0; 136 137 if (!static_branch_likely(&sched_smt_present)) 138 return -ENODEV; 139 140 BUILD_BUG_ON(PR_SCHED_CORE_SCOPE_THREAD != PIDTYPE_PID); 141 BUILD_BUG_ON(PR_SCHED_CORE_SCOPE_THREAD_GROUP != PIDTYPE_TGID); 142 BUILD_BUG_ON(PR_SCHED_CORE_SCOPE_PROCESS_GROUP != PIDTYPE_PGID); 143 144 if (type > PIDTYPE_PGID || cmd >= PR_SCHED_CORE_MAX || pid < 0 || 145 (cmd != PR_SCHED_CORE_GET && uaddr)) 146 return -EINVAL; 147 148 rcu_read_lock(); 149 if (pid == 0) { 150 task = current; 151 } else { 152 task = find_task_by_vpid(pid); 153 if (!task) { 154 rcu_read_unlock(); 155 return -ESRCH; 156 } 157 } 158 get_task_struct(task); 159 rcu_read_unlock(); 160 161 /* 162 * Check if this process has the right to modify the specified 163 * process. Use the regular "ptrace_may_access()" checks. 164 */ 165 if (!ptrace_may_access(task, PTRACE_MODE_READ_REALCREDS)) { 166 err = -EPERM; 167 goto out; 168 } 169 170 switch (cmd) { 171 case PR_SCHED_CORE_GET: 172 if (type != PIDTYPE_PID || uaddr & 7) { 173 err = -EINVAL; 174 goto out; 175 } 176 cookie = sched_core_clone_cookie(task); 177 if (cookie) { 178 /* XXX improve ? */ 179 ptr_to_hashval((void *)cookie, &id); 180 } 181 err = put_user(id, (u64 __user *)uaddr); 182 goto out; 183 184 case PR_SCHED_CORE_CREATE: 185 cookie = sched_core_alloc_cookie(); 186 if (!cookie) { 187 err = -ENOMEM; 188 goto out; 189 } 190 break; 191 192 case PR_SCHED_CORE_SHARE_TO: 193 cookie = sched_core_clone_cookie(current); 194 break; 195 196 case PR_SCHED_CORE_SHARE_FROM: 197 if (type != PIDTYPE_PID) { 198 err = -EINVAL; 199 goto out; 200 } 201 cookie = sched_core_clone_cookie(task); 202 __sched_core_set(current, cookie); 203 goto out; 204 205 default: 206 err = -EINVAL; 207 goto out; 208 } 209 210 if (type == PIDTYPE_PID) { 211 __sched_core_set(task, cookie); 212 goto out; 213 } 214 215 read_lock(&tasklist_lock); 216 grp = task_pid_type(task, type); 217 218 do_each_pid_thread(grp, type, p) { 219 if (!ptrace_may_access(p, PTRACE_MODE_READ_REALCREDS)) { 220 err = -EPERM; 221 goto out_tasklist; 222 } 223 } while_each_pid_thread(grp, type, p); 224 225 do_each_pid_thread(grp, type, p) { 226 __sched_core_set(p, cookie); 227 } while_each_pid_thread(grp, type, p); 228 out_tasklist: 229 read_unlock(&tasklist_lock); 230 231 out: 232 sched_core_put_cookie(cookie); 233 put_task_struct(task); 234 return err; 235 } 236 237 #ifdef CONFIG_SCHEDSTATS 238 239 /* REQUIRES: rq->core's clock recently updated. */ 240 void __sched_core_account_forceidle(struct rq *rq) 241 { 242 const struct cpumask *smt_mask = cpu_smt_mask(cpu_of(rq)); 243 u64 delta, now = rq_clock(rq->core); 244 struct rq *rq_i; 245 struct task_struct *p; 246 int i; 247 248 lockdep_assert_rq_held(rq); 249 250 WARN_ON_ONCE(!rq->core->core_forceidle_count); 251 252 if (rq->core->core_forceidle_start == 0) 253 return; 254 255 delta = now - rq->core->core_forceidle_start; 256 if (unlikely((s64)delta <= 0)) 257 return; 258 259 rq->core->core_forceidle_start = now; 260 261 if (WARN_ON_ONCE(!rq->core->core_forceidle_occupation)) { 262 /* can't be forced idle without a running task */ 263 } else if (rq->core->core_forceidle_count > 1 || 264 rq->core->core_forceidle_occupation > 1) { 265 /* 266 * For larger SMT configurations, we need to scale the charged 267 * forced idle amount since there can be more than one forced 268 * idle sibling and more than one running cookied task. 269 */ 270 delta *= rq->core->core_forceidle_count; 271 delta = div_u64(delta, rq->core->core_forceidle_occupation); 272 } 273 274 for_each_cpu(i, smt_mask) { 275 rq_i = cpu_rq(i); 276 p = rq_i->core_pick ?: rq_i->curr; 277 278 if (p == rq_i->idle) 279 continue; 280 281 /* 282 * Note: this will account forceidle to the current CPU, even 283 * if it comes from our SMT sibling. 284 */ 285 __account_forceidle_time(p, delta); 286 } 287 } 288 289 void __sched_core_tick(struct rq *rq) 290 { 291 if (!rq->core->core_forceidle_count) 292 return; 293 294 if (rq != rq->core) 295 update_rq_clock(rq->core); 296 297 __sched_core_account_forceidle(rq); 298 } 299 300 #endif /* CONFIG_SCHEDSTATS */ 301