[PATCH] m68k: introduce task_thread_info
new helper - task_thread_info(task). On platforms that have thread_info allocated separately (i.e. in default case) it simply returns task->thread_info. m68k wants (and for good reasons) to embed its thread_info into task_struct. So it will (in later patch) have task_thread_info() of its own. For now we just add a macro for generic case and convert existing instances of its body in core kernel to uses of new macro. Obviously safe - all normal architectures get the same preprocessor output they used to get. Signed-off-by: Al Viro <viro@parcelfarce.linux.theplanet.co.uk> Signed-off-by: Roman Zippel <zippel@linux-m68k.org> Signed-off-by: Andrew Morton <akpm@osdl.org> Signed-off-by: Linus Torvalds <torvalds@osdl.org>
This commit is contained in:
parent
7feacd5334
commit
a1261f5461
@ -1233,32 +1233,34 @@ static inline void task_unlock(struct task_struct *p)
|
||||
spin_unlock(&p->alloc_lock);
|
||||
}
|
||||
|
||||
#define task_thread_info(task) (task)->thread_info
|
||||
|
||||
/* set thread flags in other task's structures
|
||||
* - see asm/thread_info.h for TIF_xxxx flags available
|
||||
*/
|
||||
static inline void set_tsk_thread_flag(struct task_struct *tsk, int flag)
|
||||
{
|
||||
set_ti_thread_flag(tsk->thread_info,flag);
|
||||
set_ti_thread_flag(task_thread_info(tsk), flag);
|
||||
}
|
||||
|
||||
static inline void clear_tsk_thread_flag(struct task_struct *tsk, int flag)
|
||||
{
|
||||
clear_ti_thread_flag(tsk->thread_info,flag);
|
||||
clear_ti_thread_flag(task_thread_info(tsk), flag);
|
||||
}
|
||||
|
||||
static inline int test_and_set_tsk_thread_flag(struct task_struct *tsk, int flag)
|
||||
{
|
||||
return test_and_set_ti_thread_flag(tsk->thread_info,flag);
|
||||
return test_and_set_ti_thread_flag(task_thread_info(tsk), flag);
|
||||
}
|
||||
|
||||
static inline int test_and_clear_tsk_thread_flag(struct task_struct *tsk, int flag)
|
||||
{
|
||||
return test_and_clear_ti_thread_flag(tsk->thread_info,flag);
|
||||
return test_and_clear_ti_thread_flag(task_thread_info(tsk), flag);
|
||||
}
|
||||
|
||||
static inline int test_tsk_thread_flag(struct task_struct *tsk, int flag)
|
||||
{
|
||||
return test_ti_thread_flag(tsk->thread_info,flag);
|
||||
return test_ti_thread_flag(task_thread_info(tsk), flag);
|
||||
}
|
||||
|
||||
static inline void set_tsk_need_resched(struct task_struct *tsk)
|
||||
@ -1329,12 +1331,12 @@ extern void signal_wake_up(struct task_struct *t, int resume_stopped);
|
||||
|
||||
static inline unsigned int task_cpu(const struct task_struct *p)
|
||||
{
|
||||
return p->thread_info->cpu;
|
||||
return task_thread_info(p)->cpu;
|
||||
}
|
||||
|
||||
static inline void set_task_cpu(struct task_struct *p, unsigned int cpu)
|
||||
{
|
||||
p->thread_info->cpu = cpu;
|
||||
task_thread_info(p)->cpu = cpu;
|
||||
}
|
||||
|
||||
#else
|
||||
|
@ -859,7 +859,7 @@ fastcall NORET_TYPE void do_exit(long code)
|
||||
if (group_dead && tsk->signal->leader)
|
||||
disassociate_ctty(1);
|
||||
|
||||
module_put(tsk->thread_info->exec_domain->module);
|
||||
module_put(task_thread_info(tsk)->exec_domain->module);
|
||||
if (tsk->binfmt)
|
||||
module_put(tsk->binfmt->module);
|
||||
|
||||
|
@ -919,7 +919,7 @@ static task_t *copy_process(unsigned long clone_flags,
|
||||
if (nr_threads >= max_threads)
|
||||
goto bad_fork_cleanup_count;
|
||||
|
||||
if (!try_module_get(p->thread_info->exec_domain->module))
|
||||
if (!try_module_get(task_thread_info(p)->exec_domain->module))
|
||||
goto bad_fork_cleanup_count;
|
||||
|
||||
if (p->binfmt && !try_module_get(p->binfmt->module))
|
||||
@ -1180,7 +1180,7 @@ bad_fork_cleanup:
|
||||
if (p->binfmt)
|
||||
module_put(p->binfmt->module);
|
||||
bad_fork_cleanup_put_domain:
|
||||
module_put(p->thread_info->exec_domain->module);
|
||||
module_put(task_thread_info(p)->exec_domain->module);
|
||||
bad_fork_cleanup_count:
|
||||
put_group_info(p->group_info);
|
||||
atomic_dec(&p->user->processes);
|
||||
|
@ -1437,7 +1437,7 @@ void fastcall sched_fork(task_t *p, int clone_flags)
|
||||
#endif
|
||||
#ifdef CONFIG_PREEMPT
|
||||
/* Want to start with kernel preemption disabled. */
|
||||
p->thread_info->preempt_count = 1;
|
||||
task_thread_info(p)->preempt_count = 1;
|
||||
#endif
|
||||
/*
|
||||
* Share the timeslice between parent and child, thus the
|
||||
@ -4410,9 +4410,9 @@ void __devinit init_idle(task_t *idle, int cpu)
|
||||
|
||||
/* Set the preempt count _outside_ the spinlocks! */
|
||||
#if defined(CONFIG_PREEMPT) && !defined(CONFIG_PREEMPT_BKL)
|
||||
idle->thread_info->preempt_count = (idle->lock_depth >= 0);
|
||||
task_thread_info(idle)->preempt_count = (idle->lock_depth >= 0);
|
||||
#else
|
||||
idle->thread_info->preempt_count = 0;
|
||||
task_thread_info(idle)->preempt_count = 0;
|
||||
#endif
|
||||
}
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user