Branch data Line data Source code
1 : : // SPDX-License-Identifier: GPL-2.0-only
2 : : /* Kernel thread helper functions.
3 : : * Copyright (C) 2004 IBM Corporation, Rusty Russell.
4 : : *
5 : : * Creation is done via kthreadd, so that we get a clean environment
6 : : * even if we're invoked from userspace (think modprobe, hotplug cpu,
7 : : * etc.).
8 : : */
9 : : #include <uapi/linux/sched/types.h>
10 : : #include <linux/sched.h>
11 : : #include <linux/sched/task.h>
12 : : #include <linux/kthread.h>
13 : : #include <linux/completion.h>
14 : : #include <linux/err.h>
15 : : #include <linux/cgroup.h>
16 : : #include <linux/cpuset.h>
17 : : #include <linux/unistd.h>
18 : : #include <linux/file.h>
19 : : #include <linux/export.h>
20 : : #include <linux/mutex.h>
21 : : #include <linux/slab.h>
22 : : #include <linux/freezer.h>
23 : : #include <linux/ptrace.h>
24 : : #include <linux/uaccess.h>
25 : : #include <linux/numa.h>
26 : : #include <trace/events/sched.h>
27 : :
28 : : static DEFINE_SPINLOCK(kthread_create_lock);
29 : : static LIST_HEAD(kthread_create_list);
30 : : struct task_struct *kthreadd_task;
31 : :
32 : : struct kthread_create_info
33 : : {
34 : : /* Information passed to kthread() from kthreadd. */
35 : : int (*threadfn)(void *data);
36 : : void *data;
37 : : int node;
38 : :
39 : : /* Result passed back to kthread_create() from kthreadd. */
40 : : struct task_struct *result;
41 : : struct completion *done;
42 : :
43 : : struct list_head list;
44 : : };
45 : :
46 : : struct kthread {
47 : : unsigned long flags;
48 : : unsigned int cpu;
49 : : void *data;
50 : : struct completion parked;
51 : : struct completion exited;
52 : : #ifdef CONFIG_BLK_CGROUP
53 : : struct cgroup_subsys_state *blkcg_css;
54 : : #endif
55 : : };
56 : :
57 : : enum KTHREAD_BITS {
58 : : KTHREAD_IS_PER_CPU = 0,
59 : : KTHREAD_SHOULD_STOP,
60 : : KTHREAD_SHOULD_PARK,
61 : : };
62 : :
63 : : static inline void set_kthread_struct(void *kthread)
64 : : {
65 : : /*
66 : : * We abuse ->set_child_tid to avoid the new member and because it
67 : : * can't be wrongly copied by copy_process(). We also rely on fact
68 : : * that the caller can't exec, so PF_KTHREAD can't be cleared.
69 : : */
70 : 3 : current->set_child_tid = (__force void __user *)kthread;
71 : : }
72 : :
73 : 3 : static inline struct kthread *to_kthread(struct task_struct *k)
74 : : {
75 : 3 : WARN_ON(!(k->flags & PF_KTHREAD));
76 : 3 : return (__force void *)k->set_child_tid;
77 : : }
78 : :
79 : 3 : void free_kthread_struct(struct task_struct *k)
80 : : {
81 : : struct kthread *kthread;
82 : :
83 : : /*
84 : : * Can be NULL if this kthread was created by kernel_thread()
85 : : * or if kmalloc() in kthread() failed.
86 : : */
87 : 3 : kthread = to_kthread(k);
88 : : #ifdef CONFIG_BLK_CGROUP
89 : 3 : WARN_ON_ONCE(kthread && kthread->blkcg_css);
90 : : #endif
91 : 3 : kfree(kthread);
92 : 3 : }
93 : :
94 : : /**
95 : : * kthread_should_stop - should this kthread return now?
96 : : *
97 : : * When someone calls kthread_stop() on your kthread, it will be woken
98 : : * and this will return true. You should then return, and your return
99 : : * value will be passed through to kthread_stop().
100 : : */
101 : 3 : bool kthread_should_stop(void)
102 : : {
103 : 3 : return test_bit(KTHREAD_SHOULD_STOP, &to_kthread(current)->flags);
104 : : }
105 : : EXPORT_SYMBOL(kthread_should_stop);
106 : :
107 : 3 : bool __kthread_should_park(struct task_struct *k)
108 : : {
109 : 3 : return test_bit(KTHREAD_SHOULD_PARK, &to_kthread(k)->flags);
110 : : }
111 : : EXPORT_SYMBOL_GPL(__kthread_should_park);
112 : :
113 : : /**
114 : : * kthread_should_park - should this kthread park now?
115 : : *
116 : : * When someone calls kthread_park() on your kthread, it will be woken
117 : : * and this will return true. You should then do the necessary
118 : : * cleanup and call kthread_parkme()
119 : : *
120 : : * Similar to kthread_should_stop(), but this keeps the thread alive
121 : : * and in a park position. kthread_unpark() "restarts" the thread and
122 : : * calls the thread function again.
123 : : */
124 : 3 : bool kthread_should_park(void)
125 : : {
126 : 3 : return __kthread_should_park(current);
127 : : }
128 : : EXPORT_SYMBOL_GPL(kthread_should_park);
129 : :
130 : : /**
131 : : * kthread_freezable_should_stop - should this freezable kthread return now?
132 : : * @was_frozen: optional out parameter, indicates whether %current was frozen
133 : : *
134 : : * kthread_should_stop() for freezable kthreads, which will enter
135 : : * refrigerator if necessary. This function is safe from kthread_stop() /
136 : : * freezer deadlock and freezable kthreads should use this function instead
137 : : * of calling try_to_freeze() directly.
138 : : */
139 : 0 : bool kthread_freezable_should_stop(bool *was_frozen)
140 : : {
141 : : bool frozen = false;
142 : :
143 : 0 : might_sleep();
144 : :
145 : 0 : if (unlikely(freezing(current)))
146 : 0 : frozen = __refrigerator(true);
147 : :
148 : 0 : if (was_frozen)
149 : 0 : *was_frozen = frozen;
150 : :
151 : 0 : return kthread_should_stop();
152 : : }
153 : : EXPORT_SYMBOL_GPL(kthread_freezable_should_stop);
154 : :
155 : : /**
156 : : * kthread_data - return data value specified on kthread creation
157 : : * @task: kthread task in question
158 : : *
159 : : * Return the data value specified when kthread @task was created.
160 : : * The caller is responsible for ensuring the validity of @task when
161 : : * calling this function.
162 : : */
163 : 3 : void *kthread_data(struct task_struct *task)
164 : : {
165 : 3 : return to_kthread(task)->data;
166 : : }
167 : :
168 : : /**
169 : : * kthread_probe_data - speculative version of kthread_data()
170 : : * @task: possible kthread task in question
171 : : *
172 : : * @task could be a kthread task. Return the data value specified when it
173 : : * was created if accessible. If @task isn't a kthread task or its data is
174 : : * inaccessible for any reason, %NULL is returned. This function requires
175 : : * that @task itself is safe to dereference.
176 : : */
177 : 1 : void *kthread_probe_data(struct task_struct *task)
178 : : {
179 : 1 : struct kthread *kthread = to_kthread(task);
180 : 1 : void *data = NULL;
181 : :
182 : 1 : probe_kernel_read(&data, &kthread->data, sizeof(data));
183 : 1 : return data;
184 : : }
185 : :
186 : 3 : static void __kthread_parkme(struct kthread *self)
187 : : {
188 : : for (;;) {
189 : : /*
190 : : * TASK_PARKED is a special state; we must serialize against
191 : : * possible pending wakeups to avoid store-store collisions on
192 : : * task->state.
193 : : *
194 : : * Such a collision might possibly result in the task state
195 : : * changin from TASK_PARKED and us failing the
196 : : * wait_task_inactive() in kthread_park().
197 : : */
198 : 3 : set_special_state(TASK_PARKED);
199 : 3 : if (!test_bit(KTHREAD_SHOULD_PARK, &self->flags))
200 : : break;
201 : :
202 : : /*
203 : : * Thread is going to call schedule(), do not preempt it,
204 : : * or the caller of kthread_park() may spend more time in
205 : : * wait_task_inactive().
206 : : */
207 : 3 : preempt_disable();
208 : 3 : complete(&self->parked);
209 : 3 : schedule_preempt_disabled();
210 : 3 : preempt_enable();
211 : 3 : }
212 : 3 : __set_current_state(TASK_RUNNING);
213 : 3 : }
214 : :
215 : 0 : void kthread_parkme(void)
216 : : {
217 : 0 : __kthread_parkme(to_kthread(current));
218 : 0 : }
219 : : EXPORT_SYMBOL_GPL(kthread_parkme);
220 : :
221 : 3 : static int kthread(void *_create)
222 : : {
223 : : /* Copy data: it's on kthread's stack */
224 : : struct kthread_create_info *create = _create;
225 : 3 : int (*threadfn)(void *data) = create->threadfn;
226 : 3 : void *data = create->data;
227 : : struct completion *done;
228 : : struct kthread *self;
229 : : int ret;
230 : :
231 : 3 : self = kzalloc(sizeof(*self), GFP_KERNEL);
232 : : set_kthread_struct(self);
233 : :
234 : : /* If user was SIGKILLed, I release the structure. */
235 : 3 : done = xchg(&create->done, NULL);
236 : 3 : if (!done) {
237 : 0 : kfree(create);
238 : 0 : do_exit(-EINTR);
239 : : }
240 : :
241 : 3 : if (!self) {
242 : 0 : create->result = ERR_PTR(-ENOMEM);
243 : 0 : complete(done);
244 : 0 : do_exit(-ENOMEM);
245 : : }
246 : :
247 : 3 : self->data = data;
248 : : init_completion(&self->exited);
249 : : init_completion(&self->parked);
250 : 3 : current->vfork_done = &self->exited;
251 : :
252 : : /* OK, tell user we're spawned, wait for stop or wakeup */
253 : 3 : __set_current_state(TASK_UNINTERRUPTIBLE);
254 : 3 : create->result = current;
255 : : /*
256 : : * Thread is going to call schedule(), do not preempt it,
257 : : * or the creator may spend more time in wait_task_inactive().
258 : : */
259 : 3 : preempt_disable();
260 : 3 : complete(done);
261 : 3 : schedule_preempt_disabled();
262 : 3 : preempt_enable();
263 : :
264 : : ret = -EINTR;
265 : 3 : if (!test_bit(KTHREAD_SHOULD_STOP, &self->flags)) {
266 : : cgroup_kthread_ready();
267 : 3 : __kthread_parkme(self);
268 : 3 : ret = threadfn(data);
269 : : }
270 : 2 : do_exit(ret);
271 : : }
272 : :
273 : : /* called from do_fork() to get node information for about to be created task */
274 : 3 : int tsk_fork_get_node(struct task_struct *tsk)
275 : : {
276 : : #ifdef CONFIG_NUMA
277 : : if (tsk == kthreadd_task)
278 : : return tsk->pref_node_fork;
279 : : #endif
280 : 3 : return NUMA_NO_NODE;
281 : : }
282 : :
283 : 3 : static void create_kthread(struct kthread_create_info *create)
284 : : {
285 : : int pid;
286 : :
287 : : #ifdef CONFIG_NUMA
288 : : current->pref_node_fork = create->node;
289 : : #endif
290 : : /* We want our own signal handler (we take no signals by default). */
291 : 3 : pid = kernel_thread(kthread, create, CLONE_FS | CLONE_FILES | SIGCHLD);
292 : 3 : if (pid < 0) {
293 : : /* If user was SIGKILLed, I release the structure. */
294 : 0 : struct completion *done = xchg(&create->done, NULL);
295 : :
296 : 0 : if (!done) {
297 : 0 : kfree(create);
298 : 3 : return;
299 : : }
300 : 0 : create->result = ERR_PTR(pid);
301 : 0 : complete(done);
302 : : }
303 : : }
304 : :
305 : : static __printf(4, 0)
306 : 3 : struct task_struct *__kthread_create_on_node(int (*threadfn)(void *data),
307 : : void *data, int node,
308 : : const char namefmt[],
309 : : va_list args)
310 : : {
311 : 3 : DECLARE_COMPLETION_ONSTACK(done);
312 : : struct task_struct *task;
313 : : struct kthread_create_info *create = kmalloc(sizeof(*create),
314 : : GFP_KERNEL);
315 : :
316 : 3 : if (!create)
317 : : return ERR_PTR(-ENOMEM);
318 : 3 : create->threadfn = threadfn;
319 : 3 : create->data = data;
320 : 3 : create->node = node;
321 : 3 : create->done = &done;
322 : :
323 : : spin_lock(&kthread_create_lock);
324 : 3 : list_add_tail(&create->list, &kthread_create_list);
325 : : spin_unlock(&kthread_create_lock);
326 : :
327 : 3 : wake_up_process(kthreadd_task);
328 : : /*
329 : : * Wait for completion in killable state, for I might be chosen by
330 : : * the OOM killer while kthreadd is trying to allocate memory for
331 : : * new kernel thread.
332 : : */
333 : 3 : if (unlikely(wait_for_completion_killable(&done))) {
334 : : /*
335 : : * If I was SIGKILLed before kthreadd (or new kernel thread)
336 : : * calls complete(), leave the cleanup of this structure to
337 : : * that thread.
338 : : */
339 : 0 : if (xchg(&create->done, NULL))
340 : : return ERR_PTR(-EINTR);
341 : : /*
342 : : * kthreadd (or new kernel thread) will call complete()
343 : : * shortly.
344 : : */
345 : 0 : wait_for_completion(&done);
346 : : }
347 : 3 : task = create->result;
348 : 3 : if (!IS_ERR(task)) {
349 : : static const struct sched_param param = { .sched_priority = 0 };
350 : : char name[TASK_COMM_LEN];
351 : :
352 : : /*
353 : : * task is already visible to other tasks, so updating
354 : : * COMM must be protected.
355 : : */
356 : 3 : vsnprintf(name, sizeof(name), namefmt, args);
357 : : set_task_comm(task, name);
358 : : /*
359 : : * root may have changed our (kthreadd's) priority or CPU mask.
360 : : * The kernel thread should not inherit these properties.
361 : : */
362 : 3 : sched_setscheduler_nocheck(task, SCHED_NORMAL, ¶m);
363 : 3 : set_cpus_allowed_ptr(task, cpu_all_mask);
364 : : }
365 : 3 : kfree(create);
366 : 3 : return task;
367 : : }
368 : :
369 : : /**
370 : : * kthread_create_on_node - create a kthread.
371 : : * @threadfn: the function to run until signal_pending(current).
372 : : * @data: data ptr for @threadfn.
373 : : * @node: task and thread structures for the thread are allocated on this node
374 : : * @namefmt: printf-style name for the thread.
375 : : *
376 : : * Description: This helper function creates and names a kernel
377 : : * thread. The thread will be stopped: use wake_up_process() to start
378 : : * it. See also kthread_run(). The new thread has SCHED_NORMAL policy and
379 : : * is affine to all CPUs.
380 : : *
381 : : * If thread is going to be bound on a particular cpu, give its node
382 : : * in @node, to get NUMA affinity for kthread stack, or else give NUMA_NO_NODE.
383 : : * When woken, the thread will run @threadfn() with @data as its
384 : : * argument. @threadfn() can either call do_exit() directly if it is a
385 : : * standalone thread for which no one will call kthread_stop(), or
386 : : * return when 'kthread_should_stop()' is true (which means
387 : : * kthread_stop() has been called). The return value should be zero
388 : : * or a negative error number; it will be passed to kthread_stop().
389 : : *
390 : : * Returns a task_struct or ERR_PTR(-ENOMEM) or ERR_PTR(-EINTR).
391 : : */
392 : 3 : struct task_struct *kthread_create_on_node(int (*threadfn)(void *data),
393 : : void *data, int node,
394 : : const char namefmt[],
395 : : ...)
396 : : {
397 : : struct task_struct *task;
398 : : va_list args;
399 : :
400 : 3 : va_start(args, namefmt);
401 : 3 : task = __kthread_create_on_node(threadfn, data, node, namefmt, args);
402 : 3 : va_end(args);
403 : :
404 : 3 : return task;
405 : : }
406 : : EXPORT_SYMBOL(kthread_create_on_node);
407 : :
408 : 3 : static void __kthread_bind_mask(struct task_struct *p, const struct cpumask *mask, long state)
409 : : {
410 : : unsigned long flags;
411 : :
412 : 3 : if (!wait_task_inactive(p, state)) {
413 : 0 : WARN_ON(1);
414 : 3 : return;
415 : : }
416 : :
417 : : /* It's safe because the task is inactive. */
418 : 3 : raw_spin_lock_irqsave(&p->pi_lock, flags);
419 : 3 : do_set_cpus_allowed(p, mask);
420 : 3 : p->flags |= PF_NO_SETAFFINITY;
421 : 3 : raw_spin_unlock_irqrestore(&p->pi_lock, flags);
422 : : }
423 : :
424 : 3 : static void __kthread_bind(struct task_struct *p, unsigned int cpu, long state)
425 : : {
426 : 3 : __kthread_bind_mask(p, cpumask_of(cpu), state);
427 : 3 : }
428 : :
429 : 3 : void kthread_bind_mask(struct task_struct *p, const struct cpumask *mask)
430 : : {
431 : 3 : __kthread_bind_mask(p, mask, TASK_UNINTERRUPTIBLE);
432 : 3 : }
433 : :
434 : : /**
435 : : * kthread_bind - bind a just-created kthread to a cpu.
436 : : * @p: thread created by kthread_create().
437 : : * @cpu: cpu (might not be online, must be possible) for @k to run on.
438 : : *
439 : : * Description: This function is equivalent to set_cpus_allowed(),
440 : : * except that @cpu doesn't need to be online, and the thread must be
441 : : * stopped (i.e., just returned from kthread_create()).
442 : : */
443 : 0 : void kthread_bind(struct task_struct *p, unsigned int cpu)
444 : : {
445 : 3 : __kthread_bind(p, cpu, TASK_UNINTERRUPTIBLE);
446 : 0 : }
447 : : EXPORT_SYMBOL(kthread_bind);
448 : :
449 : : /**
450 : : * kthread_create_on_cpu - Create a cpu bound kthread
451 : : * @threadfn: the function to run until signal_pending(current).
452 : : * @data: data ptr for @threadfn.
453 : : * @cpu: The cpu on which the thread should be bound,
454 : : * @namefmt: printf-style name for the thread. Format is restricted
455 : : * to "name.*%u". Code fills in cpu number.
456 : : *
457 : : * Description: This helper function creates and names a kernel thread
458 : : * The thread will be woken and put into park mode.
459 : : */
460 : 3 : struct task_struct *kthread_create_on_cpu(int (*threadfn)(void *data),
461 : : void *data, unsigned int cpu,
462 : : const char *namefmt)
463 : : {
464 : : struct task_struct *p;
465 : :
466 : 3 : p = kthread_create_on_node(threadfn, data, cpu_to_node(cpu), namefmt,
467 : : cpu);
468 : 3 : if (IS_ERR(p))
469 : : return p;
470 : : kthread_bind(p, cpu);
471 : : /* CPU hotplug need to bind once again when unparking the thread. */
472 : 3 : set_bit(KTHREAD_IS_PER_CPU, &to_kthread(p)->flags);
473 : 3 : to_kthread(p)->cpu = cpu;
474 : 3 : return p;
475 : : }
476 : :
477 : : /**
478 : : * kthread_unpark - unpark a thread created by kthread_create().
479 : : * @k: thread created by kthread_create().
480 : : *
481 : : * Sets kthread_should_park() for @k to return false, wakes it, and
482 : : * waits for it to return. If the thread is marked percpu then its
483 : : * bound to the cpu again.
484 : : */
485 : 3 : void kthread_unpark(struct task_struct *k)
486 : : {
487 : 3 : struct kthread *kthread = to_kthread(k);
488 : :
489 : : /*
490 : : * Newly created kthread was parked when the CPU was offline.
491 : : * The binding was lost and we need to set it again.
492 : : */
493 : 3 : if (test_bit(KTHREAD_IS_PER_CPU, &kthread->flags))
494 : 3 : __kthread_bind(k, kthread->cpu, TASK_PARKED);
495 : :
496 : 3 : clear_bit(KTHREAD_SHOULD_PARK, &kthread->flags);
497 : : /*
498 : : * __kthread_parkme() will either see !SHOULD_PARK or get the wakeup.
499 : : */
500 : 3 : wake_up_state(k, TASK_PARKED);
501 : 3 : }
502 : : EXPORT_SYMBOL_GPL(kthread_unpark);
503 : :
504 : : /**
505 : : * kthread_park - park a thread created by kthread_create().
506 : : * @k: thread created by kthread_create().
507 : : *
508 : : * Sets kthread_should_park() for @k to return true, wakes it, and
509 : : * waits for it to return. This can also be called after kthread_create()
510 : : * instead of calling wake_up_process(): the thread will park without
511 : : * calling threadfn().
512 : : *
513 : : * Returns 0 if the thread is parked, -ENOSYS if the thread exited.
514 : : * If called by the kthread itself just the park bit is set.
515 : : */
516 : 3 : int kthread_park(struct task_struct *k)
517 : : {
518 : 3 : struct kthread *kthread = to_kthread(k);
519 : :
520 : 3 : if (WARN_ON(k->flags & PF_EXITING))
521 : : return -ENOSYS;
522 : :
523 : 3 : if (WARN_ON_ONCE(test_bit(KTHREAD_SHOULD_PARK, &kthread->flags)))
524 : : return -EBUSY;
525 : :
526 : 3 : set_bit(KTHREAD_SHOULD_PARK, &kthread->flags);
527 : 3 : if (k != current) {
528 : 3 : wake_up_process(k);
529 : : /*
530 : : * Wait for __kthread_parkme() to complete(), this means we
531 : : * _will_ have TASK_PARKED and are about to call schedule().
532 : : */
533 : 3 : wait_for_completion(&kthread->parked);
534 : : /*
535 : : * Now wait for that schedule() to complete and the task to
536 : : * get scheduled out.
537 : : */
538 : 3 : WARN_ON_ONCE(!wait_task_inactive(k, TASK_PARKED));
539 : : }
540 : :
541 : : return 0;
542 : : }
543 : : EXPORT_SYMBOL_GPL(kthread_park);
544 : :
545 : : /**
546 : : * kthread_stop - stop a thread created by kthread_create().
547 : : * @k: thread created by kthread_create().
548 : : *
549 : : * Sets kthread_should_stop() for @k to return true, wakes it, and
550 : : * waits for it to exit. This can also be called after kthread_create()
551 : : * instead of calling wake_up_process(): the thread will exit without
552 : : * calling threadfn().
553 : : *
554 : : * If threadfn() may call do_exit() itself, the caller must ensure
555 : : * task_struct can't go away.
556 : : *
557 : : * Returns the result of threadfn(), or %-EINTR if wake_up_process()
558 : : * was never called.
559 : : */
560 : 2 : int kthread_stop(struct task_struct *k)
561 : : {
562 : : struct kthread *kthread;
563 : : int ret;
564 : :
565 : 2 : trace_sched_kthread_stop(k);
566 : :
567 : : get_task_struct(k);
568 : 2 : kthread = to_kthread(k);
569 : 2 : set_bit(KTHREAD_SHOULD_STOP, &kthread->flags);
570 : 2 : kthread_unpark(k);
571 : 2 : wake_up_process(k);
572 : 2 : wait_for_completion(&kthread->exited);
573 : 2 : ret = k->exit_code;
574 : 2 : put_task_struct(k);
575 : :
576 : 2 : trace_sched_kthread_stop_ret(ret);
577 : 2 : return ret;
578 : : }
579 : : EXPORT_SYMBOL(kthread_stop);
580 : :
581 : 3 : int kthreadd(void *unused)
582 : : {
583 : 3 : struct task_struct *tsk = current;
584 : :
585 : : /* Setup a clean context for our children to inherit. */
586 : : set_task_comm(tsk, "kthreadd");
587 : 3 : ignore_signals(tsk);
588 : 3 : set_cpus_allowed_ptr(tsk, cpu_all_mask);
589 : 3 : set_mems_allowed(node_states[N_MEMORY]);
590 : :
591 : 3 : current->flags |= PF_NOFREEZE;
592 : : cgroup_init_kthreadd();
593 : :
594 : : for (;;) {
595 : 3 : set_current_state(TASK_INTERRUPTIBLE);
596 : 3 : if (list_empty(&kthread_create_list))
597 : 3 : schedule();
598 : 3 : __set_current_state(TASK_RUNNING);
599 : :
600 : : spin_lock(&kthread_create_lock);
601 : 3 : while (!list_empty(&kthread_create_list)) {
602 : : struct kthread_create_info *create;
603 : :
604 : 3 : create = list_entry(kthread_create_list.next,
605 : : struct kthread_create_info, list);
606 : 3 : list_del_init(&create->list);
607 : : spin_unlock(&kthread_create_lock);
608 : :
609 : 3 : create_kthread(create);
610 : :
611 : : spin_lock(&kthread_create_lock);
612 : : }
613 : : spin_unlock(&kthread_create_lock);
614 : : }
615 : :
616 : : return 0;
617 : : }
618 : :
619 : 0 : void __kthread_init_worker(struct kthread_worker *worker,
620 : : const char *name,
621 : : struct lock_class_key *key)
622 : : {
623 : 3 : memset(worker, 0, sizeof(struct kthread_worker));
624 : 3 : raw_spin_lock_init(&worker->lock);
625 : : lockdep_set_class_and_name(&worker->lock, key, name);
626 : 3 : INIT_LIST_HEAD(&worker->work_list);
627 : 3 : INIT_LIST_HEAD(&worker->delayed_work_list);
628 : 0 : }
629 : : EXPORT_SYMBOL_GPL(__kthread_init_worker);
630 : :
631 : : /**
632 : : * kthread_worker_fn - kthread function to process kthread_worker
633 : : * @worker_ptr: pointer to initialized kthread_worker
634 : : *
635 : : * This function implements the main cycle of kthread worker. It processes
636 : : * work_list until it is stopped with kthread_stop(). It sleeps when the queue
637 : : * is empty.
638 : : *
639 : : * The works are not allowed to keep any locks, disable preemption or interrupts
640 : : * when they finish. There is defined a safe point for freezing when one work
641 : : * finishes and before a new one is started.
642 : : *
643 : : * Also the works must not be handled by more than one worker at the same time,
644 : : * see also kthread_queue_work().
645 : : */
646 : 3 : int kthread_worker_fn(void *worker_ptr)
647 : : {
648 : : struct kthread_worker *worker = worker_ptr;
649 : : struct kthread_work *work;
650 : :
651 : : /*
652 : : * FIXME: Update the check and remove the assignment when all kthread
653 : : * worker users are created using kthread_create_worker*() functions.
654 : : */
655 : 3 : WARN_ON(worker->task && worker->task != current);
656 : 3 : worker->task = current;
657 : :
658 : 3 : if (worker->flags & KTW_FREEZABLE)
659 : 0 : set_freezable();
660 : :
661 : : repeat:
662 : 3 : set_current_state(TASK_INTERRUPTIBLE); /* mb paired w/ kthread_stop */
663 : :
664 : 3 : if (kthread_should_stop()) {
665 : 0 : __set_current_state(TASK_RUNNING);
666 : 0 : raw_spin_lock_irq(&worker->lock);
667 : 0 : worker->task = NULL;
668 : 0 : raw_spin_unlock_irq(&worker->lock);
669 : 0 : return 0;
670 : : }
671 : :
672 : : work = NULL;
673 : 3 : raw_spin_lock_irq(&worker->lock);
674 : 3 : if (!list_empty(&worker->work_list)) {
675 : 0 : work = list_first_entry(&worker->work_list,
676 : : struct kthread_work, node);
677 : 0 : list_del_init(&work->node);
678 : : }
679 : 3 : worker->current_work = work;
680 : 3 : raw_spin_unlock_irq(&worker->lock);
681 : :
682 : 3 : if (work) {
683 : 0 : __set_current_state(TASK_RUNNING);
684 : 0 : work->func(work);
685 : 3 : } else if (!freezing(current))
686 : 3 : schedule();
687 : :
688 : : try_to_freeze();
689 : 0 : cond_resched();
690 : 0 : goto repeat;
691 : : }
692 : : EXPORT_SYMBOL_GPL(kthread_worker_fn);
693 : :
694 : : static __printf(3, 0) struct kthread_worker *
695 : 3 : __kthread_create_worker(int cpu, unsigned int flags,
696 : : const char namefmt[], va_list args)
697 : : {
698 : : struct kthread_worker *worker;
699 : : struct task_struct *task;
700 : : int node = NUMA_NO_NODE;
701 : :
702 : 3 : worker = kzalloc(sizeof(*worker), GFP_KERNEL);
703 : 3 : if (!worker)
704 : : return ERR_PTR(-ENOMEM);
705 : :
706 : : kthread_init_worker(worker);
707 : :
708 : 3 : if (cpu >= 0)
709 : : node = cpu_to_node(cpu);
710 : :
711 : 3 : task = __kthread_create_on_node(kthread_worker_fn, worker,
712 : : node, namefmt, args);
713 : 3 : if (IS_ERR(task))
714 : : goto fail_task;
715 : :
716 : 3 : if (cpu >= 0)
717 : 0 : kthread_bind(task, cpu);
718 : :
719 : 3 : worker->flags = flags;
720 : 3 : worker->task = task;
721 : 3 : wake_up_process(task);
722 : 3 : return worker;
723 : :
724 : : fail_task:
725 : 0 : kfree(worker);
726 : 0 : return ERR_CAST(task);
727 : : }
728 : :
729 : : /**
730 : : * kthread_create_worker - create a kthread worker
731 : : * @flags: flags modifying the default behavior of the worker
732 : : * @namefmt: printf-style name for the kthread worker (task).
733 : : *
734 : : * Returns a pointer to the allocated worker on success, ERR_PTR(-ENOMEM)
735 : : * when the needed structures could not get allocated, and ERR_PTR(-EINTR)
736 : : * when the worker was SIGKILLed.
737 : : */
738 : : struct kthread_worker *
739 : 3 : kthread_create_worker(unsigned int flags, const char namefmt[], ...)
740 : : {
741 : : struct kthread_worker *worker;
742 : : va_list args;
743 : :
744 : 3 : va_start(args, namefmt);
745 : 3 : worker = __kthread_create_worker(-1, flags, namefmt, args);
746 : 3 : va_end(args);
747 : :
748 : 3 : return worker;
749 : : }
750 : : EXPORT_SYMBOL(kthread_create_worker);
751 : :
752 : : /**
753 : : * kthread_create_worker_on_cpu - create a kthread worker and bind it
754 : : * it to a given CPU and the associated NUMA node.
755 : : * @cpu: CPU number
756 : : * @flags: flags modifying the default behavior of the worker
757 : : * @namefmt: printf-style name for the kthread worker (task).
758 : : *
759 : : * Use a valid CPU number if you want to bind the kthread worker
760 : : * to the given CPU and the associated NUMA node.
761 : : *
762 : : * A good practice is to add the cpu number also into the worker name.
763 : : * For example, use kthread_create_worker_on_cpu(cpu, "helper/%d", cpu).
764 : : *
765 : : * Returns a pointer to the allocated worker on success, ERR_PTR(-ENOMEM)
766 : : * when the needed structures could not get allocated, and ERR_PTR(-EINTR)
767 : : * when the worker was SIGKILLed.
768 : : */
769 : : struct kthread_worker *
770 : 0 : kthread_create_worker_on_cpu(int cpu, unsigned int flags,
771 : : const char namefmt[], ...)
772 : : {
773 : : struct kthread_worker *worker;
774 : : va_list args;
775 : :
776 : 0 : va_start(args, namefmt);
777 : 0 : worker = __kthread_create_worker(cpu, flags, namefmt, args);
778 : 0 : va_end(args);
779 : :
780 : 0 : return worker;
781 : : }
782 : : EXPORT_SYMBOL(kthread_create_worker_on_cpu);
783 : :
784 : : /*
785 : : * Returns true when the work could not be queued at the moment.
786 : : * It happens when it is already pending in a worker list
787 : : * or when it is being cancelled.
788 : : */
789 : : static inline bool queuing_blocked(struct kthread_worker *worker,
790 : : struct kthread_work *work)
791 : : {
792 : : lockdep_assert_held(&worker->lock);
793 : :
794 : 0 : return !list_empty(&work->node) || work->canceling;
795 : : }
796 : :
797 : 0 : static void kthread_insert_work_sanity_check(struct kthread_worker *worker,
798 : : struct kthread_work *work)
799 : : {
800 : : lockdep_assert_held(&worker->lock);
801 : 0 : WARN_ON_ONCE(!list_empty(&work->node));
802 : : /* Do not use a work with >1 worker, see kthread_queue_work() */
803 : 0 : WARN_ON_ONCE(work->worker && work->worker != worker);
804 : 0 : }
805 : :
806 : : /* insert @work before @pos in @worker */
807 : 0 : static void kthread_insert_work(struct kthread_worker *worker,
808 : : struct kthread_work *work,
809 : : struct list_head *pos)
810 : : {
811 : 0 : kthread_insert_work_sanity_check(worker, work);
812 : :
813 : 0 : list_add_tail(&work->node, pos);
814 : 0 : work->worker = worker;
815 : 0 : if (!worker->current_work && likely(worker->task))
816 : 0 : wake_up_process(worker->task);
817 : 0 : }
818 : :
819 : : /**
820 : : * kthread_queue_work - queue a kthread_work
821 : : * @worker: target kthread_worker
822 : : * @work: kthread_work to queue
823 : : *
824 : : * Queue @work to work processor @task for async execution. @task
825 : : * must have been created with kthread_worker_create(). Returns %true
826 : : * if @work was successfully queued, %false if it was already pending.
827 : : *
828 : : * Reinitialize the work if it needs to be used by another worker.
829 : : * For example, when the worker was stopped and started again.
830 : : */
831 : 0 : bool kthread_queue_work(struct kthread_worker *worker,
832 : : struct kthread_work *work)
833 : : {
834 : : bool ret = false;
835 : : unsigned long flags;
836 : :
837 : 0 : raw_spin_lock_irqsave(&worker->lock, flags);
838 : 0 : if (!queuing_blocked(worker, work)) {
839 : 0 : kthread_insert_work(worker, work, &worker->work_list);
840 : : ret = true;
841 : : }
842 : 0 : raw_spin_unlock_irqrestore(&worker->lock, flags);
843 : 0 : return ret;
844 : : }
845 : : EXPORT_SYMBOL_GPL(kthread_queue_work);
846 : :
847 : : /**
848 : : * kthread_delayed_work_timer_fn - callback that queues the associated kthread
849 : : * delayed work when the timer expires.
850 : : * @t: pointer to the expired timer
851 : : *
852 : : * The format of the function is defined by struct timer_list.
853 : : * It should have been called from irqsafe timer with irq already off.
854 : : */
855 : 0 : void kthread_delayed_work_timer_fn(struct timer_list *t)
856 : : {
857 : : struct kthread_delayed_work *dwork = from_timer(dwork, t, timer);
858 : 0 : struct kthread_work *work = &dwork->work;
859 : 0 : struct kthread_worker *worker = work->worker;
860 : : unsigned long flags;
861 : :
862 : : /*
863 : : * This might happen when a pending work is reinitialized.
864 : : * It means that it is used a wrong way.
865 : : */
866 : 0 : if (WARN_ON_ONCE(!worker))
867 : 0 : return;
868 : :
869 : 0 : raw_spin_lock_irqsave(&worker->lock, flags);
870 : : /* Work must not be used with >1 worker, see kthread_queue_work(). */
871 : 0 : WARN_ON_ONCE(work->worker != worker);
872 : :
873 : : /* Move the work from worker->delayed_work_list. */
874 : 0 : WARN_ON_ONCE(list_empty(&work->node));
875 : : list_del_init(&work->node);
876 : 0 : kthread_insert_work(worker, work, &worker->work_list);
877 : :
878 : 0 : raw_spin_unlock_irqrestore(&worker->lock, flags);
879 : : }
880 : : EXPORT_SYMBOL(kthread_delayed_work_timer_fn);
881 : :
882 : 0 : static void __kthread_queue_delayed_work(struct kthread_worker *worker,
883 : : struct kthread_delayed_work *dwork,
884 : : unsigned long delay)
885 : : {
886 : 0 : struct timer_list *timer = &dwork->timer;
887 : 0 : struct kthread_work *work = &dwork->work;
888 : :
889 : 0 : WARN_ON_ONCE(timer->function != kthread_delayed_work_timer_fn);
890 : :
891 : : /*
892 : : * If @delay is 0, queue @dwork->work immediately. This is for
893 : : * both optimization and correctness. The earliest @timer can
894 : : * expire is on the closest next tick and delayed_work users depend
895 : : * on that there's no such delay when @delay is 0.
896 : : */
897 : 0 : if (!delay) {
898 : 0 : kthread_insert_work(worker, work, &worker->work_list);
899 : 0 : return;
900 : : }
901 : :
902 : : /* Be paranoid and try to detect possible races already now. */
903 : 0 : kthread_insert_work_sanity_check(worker, work);
904 : :
905 : 0 : list_add(&work->node, &worker->delayed_work_list);
906 : 0 : work->worker = worker;
907 : 0 : timer->expires = jiffies + delay;
908 : 0 : add_timer(timer);
909 : : }
910 : :
911 : : /**
912 : : * kthread_queue_delayed_work - queue the associated kthread work
913 : : * after a delay.
914 : : * @worker: target kthread_worker
915 : : * @dwork: kthread_delayed_work to queue
916 : : * @delay: number of jiffies to wait before queuing
917 : : *
918 : : * If the work has not been pending it starts a timer that will queue
919 : : * the work after the given @delay. If @delay is zero, it queues the
920 : : * work immediately.
921 : : *
922 : : * Return: %false if the @work has already been pending. It means that
923 : : * either the timer was running or the work was queued. It returns %true
924 : : * otherwise.
925 : : */
926 : 0 : bool kthread_queue_delayed_work(struct kthread_worker *worker,
927 : : struct kthread_delayed_work *dwork,
928 : : unsigned long delay)
929 : : {
930 : : struct kthread_work *work = &dwork->work;
931 : : unsigned long flags;
932 : : bool ret = false;
933 : :
934 : 0 : raw_spin_lock_irqsave(&worker->lock, flags);
935 : :
936 : 0 : if (!queuing_blocked(worker, work)) {
937 : 0 : __kthread_queue_delayed_work(worker, dwork, delay);
938 : : ret = true;
939 : : }
940 : :
941 : 0 : raw_spin_unlock_irqrestore(&worker->lock, flags);
942 : 0 : return ret;
943 : : }
944 : : EXPORT_SYMBOL_GPL(kthread_queue_delayed_work);
945 : :
946 : : struct kthread_flush_work {
947 : : struct kthread_work work;
948 : : struct completion done;
949 : : };
950 : :
951 : 0 : static void kthread_flush_work_fn(struct kthread_work *work)
952 : : {
953 : : struct kthread_flush_work *fwork =
954 : : container_of(work, struct kthread_flush_work, work);
955 : 0 : complete(&fwork->done);
956 : 0 : }
957 : :
958 : : /**
959 : : * kthread_flush_work - flush a kthread_work
960 : : * @work: work to flush
961 : : *
962 : : * If @work is queued or executing, wait for it to finish execution.
963 : : */
964 : 0 : void kthread_flush_work(struct kthread_work *work)
965 : : {
966 : 0 : struct kthread_flush_work fwork = {
967 : : KTHREAD_WORK_INIT(fwork.work, kthread_flush_work_fn),
968 : : COMPLETION_INITIALIZER_ONSTACK(fwork.done),
969 : : };
970 : : struct kthread_worker *worker;
971 : : bool noop = false;
972 : :
973 : 0 : worker = work->worker;
974 : 0 : if (!worker)
975 : 0 : return;
976 : :
977 : 0 : raw_spin_lock_irq(&worker->lock);
978 : : /* Work must not be used with >1 worker, see kthread_queue_work(). */
979 : 0 : WARN_ON_ONCE(work->worker != worker);
980 : :
981 : 0 : if (!list_empty(&work->node))
982 : 0 : kthread_insert_work(worker, &fwork.work, work->node.next);
983 : 0 : else if (worker->current_work == work)
984 : 0 : kthread_insert_work(worker, &fwork.work,
985 : : worker->work_list.next);
986 : : else
987 : : noop = true;
988 : :
989 : 0 : raw_spin_unlock_irq(&worker->lock);
990 : :
991 : 0 : if (!noop)
992 : 0 : wait_for_completion(&fwork.done);
993 : : }
994 : : EXPORT_SYMBOL_GPL(kthread_flush_work);
995 : :
996 : : /*
997 : : * This function removes the work from the worker queue. Also it makes sure
998 : : * that it won't get queued later via the delayed work's timer.
999 : : *
1000 : : * The work might still be in use when this function finishes. See the
1001 : : * current_work proceed by the worker.
1002 : : *
1003 : : * Return: %true if @work was pending and successfully canceled,
1004 : : * %false if @work was not pending
1005 : : */
1006 : 0 : static bool __kthread_cancel_work(struct kthread_work *work, bool is_dwork,
1007 : : unsigned long *flags)
1008 : : {
1009 : : /* Try to cancel the timer if exists. */
1010 : 0 : if (is_dwork) {
1011 : : struct kthread_delayed_work *dwork =
1012 : : container_of(work, struct kthread_delayed_work, work);
1013 : 0 : struct kthread_worker *worker = work->worker;
1014 : :
1015 : : /*
1016 : : * del_timer_sync() must be called to make sure that the timer
1017 : : * callback is not running. The lock must be temporary released
1018 : : * to avoid a deadlock with the callback. In the meantime,
1019 : : * any queuing is blocked by setting the canceling counter.
1020 : : */
1021 : 0 : work->canceling++;
1022 : 0 : raw_spin_unlock_irqrestore(&worker->lock, *flags);
1023 : 0 : del_timer_sync(&dwork->timer);
1024 : 0 : raw_spin_lock_irqsave(&worker->lock, *flags);
1025 : 0 : work->canceling--;
1026 : : }
1027 : :
1028 : : /*
1029 : : * Try to remove the work from a worker list. It might either
1030 : : * be from worker->work_list or from worker->delayed_work_list.
1031 : : */
1032 : 0 : if (!list_empty(&work->node)) {
1033 : : list_del_init(&work->node);
1034 : 0 : return true;
1035 : : }
1036 : :
1037 : : return false;
1038 : : }
1039 : :
1040 : : /**
1041 : : * kthread_mod_delayed_work - modify delay of or queue a kthread delayed work
1042 : : * @worker: kthread worker to use
1043 : : * @dwork: kthread delayed work to queue
1044 : : * @delay: number of jiffies to wait before queuing
1045 : : *
1046 : : * If @dwork is idle, equivalent to kthread_queue_delayed_work(). Otherwise,
1047 : : * modify @dwork's timer so that it expires after @delay. If @delay is zero,
1048 : : * @work is guaranteed to be queued immediately.
1049 : : *
1050 : : * Return: %true if @dwork was pending and its timer was modified,
1051 : : * %false otherwise.
1052 : : *
1053 : : * A special case is when the work is being canceled in parallel.
1054 : : * It might be caused either by the real kthread_cancel_delayed_work_sync()
1055 : : * or yet another kthread_mod_delayed_work() call. We let the other command
1056 : : * win and return %false here. The caller is supposed to synchronize these
1057 : : * operations a reasonable way.
1058 : : *
1059 : : * This function is safe to call from any context including IRQ handler.
1060 : : * See __kthread_cancel_work() and kthread_delayed_work_timer_fn()
1061 : : * for details.
1062 : : */
1063 : 0 : bool kthread_mod_delayed_work(struct kthread_worker *worker,
1064 : : struct kthread_delayed_work *dwork,
1065 : : unsigned long delay)
1066 : : {
1067 : 0 : struct kthread_work *work = &dwork->work;
1068 : : unsigned long flags;
1069 : : int ret = false;
1070 : :
1071 : 0 : raw_spin_lock_irqsave(&worker->lock, flags);
1072 : :
1073 : : /* Do not bother with canceling when never queued. */
1074 : 0 : if (!work->worker)
1075 : : goto fast_queue;
1076 : :
1077 : : /* Work must not be used with >1 worker, see kthread_queue_work() */
1078 : 0 : WARN_ON_ONCE(work->worker != worker);
1079 : :
1080 : : /* Do not fight with another command that is canceling this work. */
1081 : 0 : if (work->canceling)
1082 : : goto out;
1083 : :
1084 : 0 : ret = __kthread_cancel_work(work, true, &flags);
1085 : : fast_queue:
1086 : 0 : __kthread_queue_delayed_work(worker, dwork, delay);
1087 : : out:
1088 : 0 : raw_spin_unlock_irqrestore(&worker->lock, flags);
1089 : 0 : return ret;
1090 : : }
1091 : : EXPORT_SYMBOL_GPL(kthread_mod_delayed_work);
1092 : :
1093 : 0 : static bool __kthread_cancel_work_sync(struct kthread_work *work, bool is_dwork)
1094 : : {
1095 : 0 : struct kthread_worker *worker = work->worker;
1096 : : unsigned long flags;
1097 : : int ret = false;
1098 : :
1099 : 0 : if (!worker)
1100 : : goto out;
1101 : :
1102 : 0 : raw_spin_lock_irqsave(&worker->lock, flags);
1103 : : /* Work must not be used with >1 worker, see kthread_queue_work(). */
1104 : 0 : WARN_ON_ONCE(work->worker != worker);
1105 : :
1106 : 0 : ret = __kthread_cancel_work(work, is_dwork, &flags);
1107 : :
1108 : 0 : if (worker->current_work != work)
1109 : : goto out_fast;
1110 : :
1111 : : /*
1112 : : * The work is in progress and we need to wait with the lock released.
1113 : : * In the meantime, block any queuing by setting the canceling counter.
1114 : : */
1115 : 0 : work->canceling++;
1116 : 0 : raw_spin_unlock_irqrestore(&worker->lock, flags);
1117 : 0 : kthread_flush_work(work);
1118 : 0 : raw_spin_lock_irqsave(&worker->lock, flags);
1119 : 0 : work->canceling--;
1120 : :
1121 : : out_fast:
1122 : 0 : raw_spin_unlock_irqrestore(&worker->lock, flags);
1123 : : out:
1124 : 0 : return ret;
1125 : : }
1126 : :
1127 : : /**
1128 : : * kthread_cancel_work_sync - cancel a kthread work and wait for it to finish
1129 : : * @work: the kthread work to cancel
1130 : : *
1131 : : * Cancel @work and wait for its execution to finish. This function
1132 : : * can be used even if the work re-queues itself. On return from this
1133 : : * function, @work is guaranteed to be not pending or executing on any CPU.
1134 : : *
1135 : : * kthread_cancel_work_sync(&delayed_work->work) must not be used for
1136 : : * delayed_work's. Use kthread_cancel_delayed_work_sync() instead.
1137 : : *
1138 : : * The caller must ensure that the worker on which @work was last
1139 : : * queued can't be destroyed before this function returns.
1140 : : *
1141 : : * Return: %true if @work was pending, %false otherwise.
1142 : : */
1143 : 0 : bool kthread_cancel_work_sync(struct kthread_work *work)
1144 : : {
1145 : 0 : return __kthread_cancel_work_sync(work, false);
1146 : : }
1147 : : EXPORT_SYMBOL_GPL(kthread_cancel_work_sync);
1148 : :
1149 : : /**
1150 : : * kthread_cancel_delayed_work_sync - cancel a kthread delayed work and
1151 : : * wait for it to finish.
1152 : : * @dwork: the kthread delayed work to cancel
1153 : : *
1154 : : * This is kthread_cancel_work_sync() for delayed works.
1155 : : *
1156 : : * Return: %true if @dwork was pending, %false otherwise.
1157 : : */
1158 : 0 : bool kthread_cancel_delayed_work_sync(struct kthread_delayed_work *dwork)
1159 : : {
1160 : 0 : return __kthread_cancel_work_sync(&dwork->work, true);
1161 : : }
1162 : : EXPORT_SYMBOL_GPL(kthread_cancel_delayed_work_sync);
1163 : :
1164 : : /**
1165 : : * kthread_flush_worker - flush all current works on a kthread_worker
1166 : : * @worker: worker to flush
1167 : : *
1168 : : * Wait until all currently executing or pending works on @worker are
1169 : : * finished.
1170 : : */
1171 : 0 : void kthread_flush_worker(struct kthread_worker *worker)
1172 : : {
1173 : 0 : struct kthread_flush_work fwork = {
1174 : : KTHREAD_WORK_INIT(fwork.work, kthread_flush_work_fn),
1175 : : COMPLETION_INITIALIZER_ONSTACK(fwork.done),
1176 : : };
1177 : :
1178 : 0 : kthread_queue_work(worker, &fwork.work);
1179 : 0 : wait_for_completion(&fwork.done);
1180 : 0 : }
1181 : : EXPORT_SYMBOL_GPL(kthread_flush_worker);
1182 : :
1183 : : /**
1184 : : * kthread_destroy_worker - destroy a kthread worker
1185 : : * @worker: worker to be destroyed
1186 : : *
1187 : : * Flush and destroy @worker. The simple flush is enough because the kthread
1188 : : * worker API is used only in trivial scenarios. There are no multi-step state
1189 : : * machines needed.
1190 : : */
1191 : 0 : void kthread_destroy_worker(struct kthread_worker *worker)
1192 : : {
1193 : : struct task_struct *task;
1194 : :
1195 : 0 : task = worker->task;
1196 : 0 : if (WARN_ON(!task))
1197 : 0 : return;
1198 : :
1199 : 0 : kthread_flush_worker(worker);
1200 : 0 : kthread_stop(task);
1201 : 0 : WARN_ON(!list_empty(&worker->work_list));
1202 : 0 : kfree(worker);
1203 : : }
1204 : : EXPORT_SYMBOL(kthread_destroy_worker);
1205 : :
1206 : : #ifdef CONFIG_BLK_CGROUP
1207 : : /**
1208 : : * kthread_associate_blkcg - associate blkcg to current kthread
1209 : : * @css: the cgroup info
1210 : : *
1211 : : * Current thread must be a kthread. The thread is running jobs on behalf of
1212 : : * other threads. In some cases, we expect the jobs attach cgroup info of
1213 : : * original threads instead of that of current thread. This function stores
1214 : : * original thread's cgroup info in current kthread context for later
1215 : : * retrieval.
1216 : : */
1217 : 0 : void kthread_associate_blkcg(struct cgroup_subsys_state *css)
1218 : : {
1219 : : struct kthread *kthread;
1220 : :
1221 : 0 : if (!(current->flags & PF_KTHREAD))
1222 : : return;
1223 : 0 : kthread = to_kthread(current);
1224 : 0 : if (!kthread)
1225 : : return;
1226 : :
1227 : 0 : if (kthread->blkcg_css) {
1228 : : css_put(kthread->blkcg_css);
1229 : 0 : kthread->blkcg_css = NULL;
1230 : : }
1231 : 0 : if (css) {
1232 : : css_get(css);
1233 : 0 : kthread->blkcg_css = css;
1234 : : }
1235 : : }
1236 : : EXPORT_SYMBOL(kthread_associate_blkcg);
1237 : :
1238 : : /**
1239 : : * kthread_blkcg - get associated blkcg css of current kthread
1240 : : *
1241 : : * Current thread must be a kthread.
1242 : : */
1243 : 3 : struct cgroup_subsys_state *kthread_blkcg(void)
1244 : : {
1245 : : struct kthread *kthread;
1246 : :
1247 : 3 : if (current->flags & PF_KTHREAD) {
1248 : 3 : kthread = to_kthread(current);
1249 : 3 : if (kthread)
1250 : 3 : return kthread->blkcg_css;
1251 : : }
1252 : : return NULL;
1253 : : }
1254 : : EXPORT_SYMBOL(kthread_blkcg);
1255 : : #endif
|