mirror of
https://github.com/Fishwaldo/linux-bl808.git
synced 2025-06-17 20:25:19 +00:00
[PATCH] sched: cleanup, remove task_t, convert to struct task_struct
cleanup: remove task_t and convert all the uses to struct task_struct. I introduced it for the scheduler anno and it was a mistake. Conversion was mostly scripted, the result was reviewed and all secondary whitespace and style impact (if any) was fixed up by hand. Signed-off-by: Ingo Molnar <mingo@elte.hu> Signed-off-by: Andrew Morton <akpm@osdl.org> Signed-off-by: Linus Torvalds <torvalds@osdl.org>
This commit is contained in:
parent
48f24c4da1
commit
36c8b58689
25 changed files with 203 additions and 187 deletions
|
@ -184,11 +184,11 @@ extern unsigned long weighted_cpuload(const int cpu);
|
|||
extern rwlock_t tasklist_lock;
|
||||
extern spinlock_t mmlist_lock;
|
||||
|
||||
typedef struct task_struct task_t;
|
||||
struct task_struct;
|
||||
|
||||
extern void sched_init(void);
|
||||
extern void sched_init_smp(void);
|
||||
extern void init_idle(task_t *idle, int cpu);
|
||||
extern void init_idle(struct task_struct *idle, int cpu);
|
||||
|
||||
extern cpumask_t nohz_cpu_mask;
|
||||
|
||||
|
@ -383,7 +383,7 @@ struct signal_struct {
|
|||
wait_queue_head_t wait_chldexit; /* for wait4() */
|
||||
|
||||
/* current thread group signal load-balancing target: */
|
||||
task_t *curr_target;
|
||||
struct task_struct *curr_target;
|
||||
|
||||
/* shared signal handling: */
|
||||
struct sigpending shared_pending;
|
||||
|
@ -699,7 +699,7 @@ extern int groups_search(struct group_info *group_info, gid_t grp);
|
|||
((gi)->blocks[(i)/NGROUPS_PER_BLOCK][(i)%NGROUPS_PER_BLOCK])
|
||||
|
||||
#ifdef ARCH_HAS_PREFETCH_SWITCH_STACK
|
||||
extern void prefetch_stack(struct task_struct*);
|
||||
extern void prefetch_stack(struct task_struct *t);
|
||||
#else
|
||||
static inline void prefetch_stack(struct task_struct *t) { }
|
||||
#endif
|
||||
|
@ -1031,9 +1031,9 @@ static inline void put_task_struct(struct task_struct *t)
|
|||
#define used_math() tsk_used_math(current)
|
||||
|
||||
#ifdef CONFIG_SMP
|
||||
extern int set_cpus_allowed(task_t *p, cpumask_t new_mask);
|
||||
extern int set_cpus_allowed(struct task_struct *p, cpumask_t new_mask);
|
||||
#else
|
||||
static inline int set_cpus_allowed(task_t *p, cpumask_t new_mask)
|
||||
static inline int set_cpus_allowed(struct task_struct *p, cpumask_t new_mask)
|
||||
{
|
||||
if (!cpu_isset(0, new_mask))
|
||||
return -EINVAL;
|
||||
|
@ -1042,7 +1042,8 @@ static inline int set_cpus_allowed(task_t *p, cpumask_t new_mask)
|
|||
#endif
|
||||
|
||||
extern unsigned long long sched_clock(void);
|
||||
extern unsigned long long current_sched_time(const task_t *current_task);
|
||||
extern unsigned long long
|
||||
current_sched_time(const struct task_struct *current_task);
|
||||
|
||||
/* sched_exec is called by processes performing an exec */
|
||||
#ifdef CONFIG_SMP
|
||||
|
@ -1060,27 +1061,27 @@ static inline void idle_task_exit(void) {}
|
|||
extern void sched_idle_next(void);
|
||||
|
||||
#ifdef CONFIG_RT_MUTEXES
|
||||
extern int rt_mutex_getprio(task_t *p);
|
||||
extern void rt_mutex_setprio(task_t *p, int prio);
|
||||
extern void rt_mutex_adjust_pi(task_t *p);
|
||||
extern int rt_mutex_getprio(struct task_struct *p);
|
||||
extern void rt_mutex_setprio(struct task_struct *p, int prio);
|
||||
extern void rt_mutex_adjust_pi(struct task_struct *p);
|
||||
#else
|
||||
static inline int rt_mutex_getprio(task_t *p)
|
||||
static inline int rt_mutex_getprio(struct task_struct *p)
|
||||
{
|
||||
return p->normal_prio;
|
||||
}
|
||||
# define rt_mutex_adjust_pi(p) do { } while (0)
|
||||
#endif
|
||||
|
||||
extern void set_user_nice(task_t *p, long nice);
|
||||
extern int task_prio(const task_t *p);
|
||||
extern int task_nice(const task_t *p);
|
||||
extern int can_nice(const task_t *p, const int nice);
|
||||
extern int task_curr(const task_t *p);
|
||||
extern void set_user_nice(struct task_struct *p, long nice);
|
||||
extern int task_prio(const struct task_struct *p);
|
||||
extern int task_nice(const struct task_struct *p);
|
||||
extern int can_nice(const struct task_struct *p, const int nice);
|
||||
extern int task_curr(const struct task_struct *p);
|
||||
extern int idle_cpu(int cpu);
|
||||
extern int sched_setscheduler(struct task_struct *, int, struct sched_param *);
|
||||
extern task_t *idle_task(int cpu);
|
||||
extern task_t *curr_task(int cpu);
|
||||
extern void set_curr_task(int cpu, task_t *p);
|
||||
extern struct task_struct *idle_task(int cpu);
|
||||
extern struct task_struct *curr_task(int cpu);
|
||||
extern void set_curr_task(int cpu, struct task_struct *p);
|
||||
|
||||
void yield(void);
|
||||
|
||||
|
@ -1137,8 +1138,8 @@ extern void FASTCALL(wake_up_new_task(struct task_struct * tsk,
|
|||
#else
|
||||
static inline void kick_process(struct task_struct *tsk) { }
|
||||
#endif
|
||||
extern void FASTCALL(sched_fork(task_t * p, int clone_flags));
|
||||
extern void FASTCALL(sched_exit(task_t * p));
|
||||
extern void FASTCALL(sched_fork(struct task_struct * p, int clone_flags));
|
||||
extern void FASTCALL(sched_exit(struct task_struct * p));
|
||||
|
||||
extern int in_group_p(gid_t);
|
||||
extern int in_egroup_p(gid_t);
|
||||
|
@ -1243,17 +1244,17 @@ extern NORET_TYPE void do_group_exit(int);
|
|||
extern void daemonize(const char *, ...);
|
||||
extern int allow_signal(int);
|
||||
extern int disallow_signal(int);
|
||||
extern task_t *child_reaper;
|
||||
extern struct task_struct *child_reaper;
|
||||
|
||||
extern int do_execve(char *, char __user * __user *, char __user * __user *, struct pt_regs *);
|
||||
extern long do_fork(unsigned long, unsigned long, struct pt_regs *, unsigned long, int __user *, int __user *);
|
||||
task_t *fork_idle(int);
|
||||
struct task_struct *fork_idle(int);
|
||||
|
||||
extern void set_task_comm(struct task_struct *tsk, char *from);
|
||||
extern void get_task_comm(char *to, struct task_struct *tsk);
|
||||
|
||||
#ifdef CONFIG_SMP
|
||||
extern void wait_task_inactive(task_t * p);
|
||||
extern void wait_task_inactive(struct task_struct * p);
|
||||
#else
|
||||
#define wait_task_inactive(p) do { } while (0)
|
||||
#endif
|
||||
|
@ -1279,13 +1280,13 @@ extern void wait_task_inactive(task_t * p);
|
|||
/* de_thread depends on thread_group_leader not being a pid based check */
|
||||
#define thread_group_leader(p) (p == p->group_leader)
|
||||
|
||||
static inline task_t *next_thread(const task_t *p)
|
||||
static inline struct task_struct *next_thread(const struct task_struct *p)
|
||||
{
|
||||
return list_entry(rcu_dereference(p->thread_group.next),
|
||||
task_t, thread_group);
|
||||
struct task_struct, thread_group);
|
||||
}
|
||||
|
||||
static inline int thread_group_empty(task_t *p)
|
||||
static inline int thread_group_empty(struct task_struct *p)
|
||||
{
|
||||
return list_empty(&p->thread_group);
|
||||
}
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue