|
|
@@ -4505,7 +4505,7 @@ void show_state_filter(unsigned long state_filter)
|
|
|
" task PC stack pid father\n");
|
|
|
#endif
|
|
|
rcu_read_lock();
|
|
|
- do_each_thread(g, p) {
|
|
|
+ for_each_process_thread(g, p) {
|
|
|
/*
|
|
|
* reset the NMI-timeout, listing all files on a slow
|
|
|
* console might take a lot of time:
|
|
|
@@ -4513,7 +4513,7 @@ void show_state_filter(unsigned long state_filter)
|
|
|
touch_nmi_watchdog();
|
|
|
if (!state_filter || (p->state & state_filter))
|
|
|
sched_show_task(p);
|
|
|
- } while_each_thread(g, p);
|
|
|
+ }
|
|
|
|
|
|
touch_all_softlockup_watchdogs();
|
|
|
|
|
|
@@ -7137,7 +7137,7 @@ void normalize_rt_tasks(void)
|
|
|
struct rq *rq;
|
|
|
|
|
|
read_lock_irqsave(&tasklist_lock, flags);
|
|
|
- do_each_thread(g, p) {
|
|
|
+ for_each_process_thread(g, p) {
|
|
|
/*
|
|
|
* Only normalize user tasks:
|
|
|
*/
|
|
|
@@ -7168,8 +7168,7 @@ void normalize_rt_tasks(void)
|
|
|
|
|
|
__task_rq_unlock(rq);
|
|
|
raw_spin_unlock(&p->pi_lock);
|
|
|
- } while_each_thread(g, p);
|
|
|
-
|
|
|
+ }
|
|
|
read_unlock_irqrestore(&tasklist_lock, flags);
|
|
|
}
|
|
|
|
|
|
@@ -7357,10 +7356,10 @@ static inline int tg_has_rt_tasks(struct task_group *tg)
|
|
|
{
|
|
|
struct task_struct *g, *p;
|
|
|
|
|
|
- do_each_thread(g, p) {
|
|
|
+ for_each_process_thread(g, p) {
|
|
|
if (rt_task(p) && task_rq(p)->rt.tg == tg)
|
|
|
return 1;
|
|
|
- } while_each_thread(g, p);
|
|
|
+ }
|
|
|
|
|
|
return 0;
|
|
|
}
|